[UPDATE] Mojo Is Faster Than Rust - Mojo Explains More

Ғылым және технология

Recorded live on twitch, GET IN
/ theprimeagen
Become a backend engineer. Its my favorite site
boot.dev/?promo=PRIMEYT
This is also the best way to support me is to support yourself becoming a better backend engineer.
Modular blog article: t.co/toEC9Yotf2
By: Jack Clayton | x.com/jack_clayto?s=21&t=-sv4...
MY MAIN YT CHANNEL: Has well edited engineering videos
/ theprimeagen
Discord
/ discord
Have something for me to read or react to?: / theprimeagenreact
Kinesis Advantage 360: bit.ly/Prime-Kinesis
Hey I am sponsored by Turso, an edge database. I think they are pretty neet. Give them a try for free and if you want you can get a decent amount off (the free tier is the best (better than planetscale or any other))
turso.tech/deeznuts

Пікірлер: 692

  • @NotMarkKnopfler
    @NotMarkKnopfler3 ай бұрын

    JUNIOR MOJO DEVELOPER REQUIRED. Must have 15 years MOJO development experience. Apply within.

  • @mac.ignacio

    @mac.ignacio

    3 ай бұрын

    I had this same experience someone in LinkedIn said he has 20 years of exp in ReactJS.

  • @georgerogers1166

    @georgerogers1166

    3 ай бұрын

    @@mac.ignacio JS makes sense, but not React LOL.

  • @NotMarkKnopfler

    @NotMarkKnopfler

    3 ай бұрын

    🤣@@mac.ignacio

  • @Patterner

    @Patterner

    3 ай бұрын

    has to pay minimum wage for EXPOSURE.

  • @kirkanos771

    @kirkanos771

    3 ай бұрын

    Dude, i've got a 40yo experience of Basic. That counts !

  • @jeffreyhymas6803
    @jeffreyhymas68033 ай бұрын

    The article: *quotes Prime's own points back at him" Prime: "I totally agree"

  • @_Lumiere_

    @_Lumiere_

    3 ай бұрын

    Chadagen

  • @smthngsmthngsmthngdarkside

    @smthngsmthngsmthngdarkside

    3 ай бұрын

    Greenhairgen

  • @future62

    @future62

    3 ай бұрын

    You have to love yourself before you can love others

  • @furiosoRIOT

    @furiosoRIOT

    3 ай бұрын

    nailed it

  • @andrewdunbar828

    @andrewdunbar828

    3 ай бұрын

    Needs to add "I stand by my words" to his vocabulary. But for now he agrees with himself.

  • @vicky5124
    @vicky51243 ай бұрын

    Having done some research about the speed claim, decompiling both of the rust and mojo binaries, I saw that mojo optimized out everything, making the resulting binary call main and returning without any work; this was proven by making a program with only a main function, and running it through the same benchmark test, which lead to the exact same results. Rust does the same optimization in any version equal or newer to 1.75.0, and its execution is so much faster that it cannot be properly calculated. In older rust versions, the compiler would create the entire recursion, and the content inside the loop looks something like: allocate 336 bytes, deallocate 336 bytes, check branch condition, call recursion function; this is slower than mojo because it is actually doing work by having to call the allocator multiple times, and having to manage a branch; also note that the compiler did do TCO, shown by having the deallocation happen before the branch condition check. This is a case where the benchmark was cherry-picked to something rust did not optimize at compile time, while also lying about rust not doing TCO, as well as using non-ideomatic rust (as shown in the video, `vec![0; size]` is preferred over `Vec::with_capacity(size)`, and results in a faster execution time than mojo). If they cannot properly explain why their language has the faster execution in the benchmark, having to assume things about how rust works to make something believable, then I do not see how any claims they make will be trusted in the foreseeable future. I am willing to change my mind in the matter if the article writer is able to explain themselves about this discrepancy, but for now, here's the truth about those benchmarks.

  • @jstn8785

    @jstn8785

    3 ай бұрын

    type shit

  • @qeqsiquemechanical9041

    @qeqsiquemechanical9041

    3 ай бұрын

    thanks for the explanation, I was wondering if some compile-time shenanigans were happening

  • @deiminator2

    @deiminator2

    3 ай бұрын

    And all of that to justify an L

  • @anonymousalexander6005

    @anonymousalexander6005

    3 ай бұрын

    @@deiminator2 No engineer worth their time in dolla bills should reach for recursion where they don't need the full stack trace, because even in Mojo you cannot rely on the optimizer in all cases, that's an NP complete problem, not happening. (a side note, a surprising number of optimizations are NP complete problems, we've always just optimized for the most common case) Mojo is targeting the AI Python dev who does not think about these things, so they optimized for their target use case, it's really just that simple. Rust targets syst... well everything really. Unfortunately, neither Rust nor Python nor Mojo is going to edge C++ within the -real- technical AI space. (it's a joke, calm down, I'm a theoretician too). Also, another side note, but MLIR (IR optimized for tensor/AI use case) can be used by Rust as well, it'd not be harder than the GCC back end. I'd be concerned about the creator getting his ego all up in him and gating MLIR optimizations behind Mojo, but I *really* don't think they will. This article is probably just marketing towards those same AI Python devs who don't understand LLVM vs MLIR or SIMD in the first place.

  • @petrusboniatus

    @petrusboniatus

    3 ай бұрын

    Expected

  • @palgun.
    @palgun.3 ай бұрын

    Green hair = rust expert

  • @mattmmilli8287

    @mattmmilli8287

    3 ай бұрын

    Green is what color your white sink will turn if there is Rust in your water ! 🤔 🤔 I think you on to something

  • @avarise5607

    @avarise5607

    3 ай бұрын

    Actually patina expert

  • @desertfish74

    @desertfish74

    3 ай бұрын

    Cyan hair 😫

  • @bgill7475

    @bgill7475

    3 ай бұрын

    Pretty much. It's just called Rust dev hair.

  • @gljames24

    @gljames24

    3 ай бұрын

    ​@@mattmmilli8287That’s verdigris, not rust.

  • @DreanPetruza
    @DreanPetruza3 ай бұрын

    I'm totally learning Mojo to be able to call myself a Mojito

  • @vectoralphaAI

    @vectoralphaAI

    3 ай бұрын

    lol, but i do love Mojician much more better.

  • @Mempler
    @Mempler3 ай бұрын

    9:20 C# is in an even weirder spot, its between 1, 2 and 3 tier. It can compile to native code, it can run without a garbage collector and it can run in IL mode. its insane

  • @FireStormOOO_

    @FireStormOOO_

    3 ай бұрын

    Yeah C# is hard to classify b/c there's a project for seemingly everything. .NET can't do X...or can it?

  • @Mr.BinarySniper

    @Mr.BinarySniper

    3 ай бұрын

    a c# fanboy detected 😁

  • @Mempler

    @Mempler

    3 ай бұрын

    @@Mr.BinarySniper ? haven't written c# since 3 years, just following its progress. No need for assumptions. Its still shit since Microsoft cant get their asses together to get a fucking UI system that will not get deprecated within a week by Microsoft itself.

  • @plaintext7288

    @plaintext7288

    3 ай бұрын

    Iirc it has two entries in the first tier - one which you described + Unity C#-to-C++ so it truly is a magical language 😂

  • @Mempler

    @Mempler

    3 ай бұрын

    @@plaintext7288 i'm talking about pure C# not even unity or anything like that. since .NET 8 has AOT support which unity effectively does with their IL2CPP

  • @robertfletcher8964
    @robertfletcher89643 ай бұрын

    Mojo programmers should clearly be called Jojos

  • @Jackovasaur

    @Jackovasaur

    3 ай бұрын

    Any compiler errors return KONO DIO DA

  • @mrpr3d4t0r

    @mrpr3d4t0r

    3 ай бұрын

    @@Jackovasaur I think he meant Mojo Jojo from powerpuff girls but that still works

  • @rollbacked

    @rollbacked

    3 ай бұрын

    if they make an IDE, they should call it Dojo

  • @gritcrit4385

    @gritcrit4385

    3 ай бұрын

    if you can't find a job with it, you will be a hobo

  • @mac.ignacio

    @mac.ignacio

    3 ай бұрын

    If they make a database in Mojo, they can call it Gojo.

  • @YaroslavFedevych
    @YaroslavFedevych3 ай бұрын

    Prime is called "Rust Experts" now.

  • @stevenhe3462

    @stevenhe3462

    3 ай бұрын

    Totally disagree. Can't even read a simple missing generic argument error message Lol.

  • @YaroslavFedevych

    @YaroslavFedevych

    3 ай бұрын

    @@stevenhe3462 ​ @stevenhe3462 It's more about how "press" (or a dude/dudette with a blog) makes things up. One Primeagen is now multiple "Rust experts". This will get repeated more and more, and next thing you know people believe all kinds of stuff said by "Rust experts", very few of them actually hauling their arse to verify what was said and if it's actually true and for which values of "win" it is true. At some iteration, someone will seriously ascribe this quote to Klabnik.

  • @ITSecNEO

    @ITSecNEO

    3 ай бұрын

    ​@@stevenhe3462Now he switches to Go lol xD I call him a language hopper (similar to distro hopper)

  • @remrevo3944
    @remrevo39443 ай бұрын

    9:25 What is kind of missing in this pyramid are insane languages like haskell, that explicitly don't surface things like execution order and other details to the user and therefore are able to do aggressive optimizations. Haskell can often get within ~5% of C performance, which is kind of insane for a high-level language.

  • @rudyperalta5592

    @rudyperalta5592

    3 ай бұрын

    With only the downside that you're writing Haskell 😢

  • @funprog

    @funprog

    3 ай бұрын

    sometimes beats C in a few benchmarks...

  • @samuraijosh1595

    @samuraijosh1595

    3 ай бұрын

    ​@@rudyperalta5592depends on who you're talking to

  • @plaintext7288

    @plaintext7288

    3 ай бұрын

    ​@@funprogif something beats C in a benchmark, the C code is poorly written

  • @gfasterOS

    @gfasterOS

    3 ай бұрын

    Looking into some of the internals for Haskell is kind of crazy. The more I learn about it the less I'm surprised that it beats C in some cases. There are definitely workloads not well suited for Haskell, but its semantics allow for some very aggressive optimization that C cannot allow.

  • @magfal
    @magfal3 ай бұрын

    26:13 you typically see a chip wide downclock when running AVX instructions on a lot of chips. You also have an overhead on the loading the input/fetching the results in many cases. In my experiments I typically see a 40X improvement, not a 64X improvement but it's constantly creeping towards that 64 number by each new architecture.

  • @Turalcar

    @Turalcar

    3 ай бұрын

    IIRC, the biggest problems with AVX512 appeared when you mixed them with scalar (or less vectorized) instructions but it could've gotten better since then.

  • @LtdJorge

    @LtdJorge

    3 ай бұрын

    That’s not true anymore. It happened on a range of Intel CPU’s only, but not for newer ones. As Turalcar says, it was the mix of AVX and scalar, because AVX-512 running at those clocks was much faster than scalar at normal clocks. It was also for AVX-512 only, but AVX-512 has been expanded massively since that, thanks to Intel again 😂

  • @LtdJorge

    @LtdJorge

    3 ай бұрын

    I think it was Cloudflare that posted in their blog, btw. They were doing AVX-512 encryption at the proxies, which was slowing down all the other things the proxy does, basically all routing and HTTP.

  • @Veptis
    @Veptis3 ай бұрын

    as a AI researcher using python all the time... my compute heavy workloads are already running on C++ under the hood. dataset.map isn't the same as a for loop... and if I use a for loop or two --- it takes a few second once or twice. Sure thing, that is good enough. The developers who integrate a c library below python via cffi are the real heros. That stuff is tough - and I am also working on something like that.

  • @The_Xeos

    @The_Xeos

    3 ай бұрын

    True, I figure maybe they want to win performance over bindings overhead? Because most AI code in Python is just calling C / CUDA libs anyway

  • @SirMo

    @SirMo

    3 ай бұрын

    This is true. The reason I never found Python to be slow for most tasks is because all the libraries which do common heavy lifting tasks are written in some lower level language by folks who are experts in the domain, so the code ends up running faster than I would be able to do it myself in a lower level language anyway.

  • @AggressivesnowmaN

    @AggressivesnowmaN

    2 ай бұрын

    But you may find the need for a custom task that runs quickly. Then you may want to pick a simple / fast language like Mojo or GO

  • @AndrewMorris-wz1vq

    @AndrewMorris-wz1vq

    25 күн бұрын

    That's what mojo is all about. Making the world under the hood pythonic by building around the new compiler.

  • @remrevo3944
    @remrevo39443 ай бұрын

    40:22 Vec::new simply doesn't allocate. Unless you push to it, it will simply point to null. That's really helpful, as that way it can just simply be used in Default impls without having to do any allocations.

  • @chinoto1

    @chinoto1

    3 ай бұрын

    I found that helpful when using `std::mem::take` to get around a borrow checker issue without a needless allocation. The current borrow checker might not need a workaround for whatever it was that I was doing.

  • @decathorpe
    @decathorpe3 ай бұрын

    One thing that doesn't seem to get mentioned at all - mojo is proprietary and not open source? That alone means it's a non-starter for so many projects and use cases that I highly doubt it will reach any kind of critical mass as a language to replace and / or supplement Python / C++ / Rust in the ML space (or any other space for that matter).

  • @ThePrimeTimeagen

    @ThePrimeTimeagen

    3 ай бұрын

    Completely agree

  • @XxZeldaxXXxLinkxX

    @XxZeldaxXXxLinkxX

    3 ай бұрын

    Oh, yeah that's a deal breaker imo. But considering that most (afaik) Ai statisticians don't care too much about what lies under the hood (their code is abysmal at times), they might not really care about whether it's open source, but that also means they might not care about using MOJO either

  • @AM-yk5yd

    @AM-yk5yd

    3 ай бұрын

    I think they promised to open source it later. I'm not holding my breath. Also if it compile to executable, it's bad thing for AI. In python I can take code of llama and hack around it. When I download LLM models I always review implementation(as most of them are (a) copies of llama, (b) small), if models were delivered as exe it would be terrible: I run linux, many AI researchers run linux, but many LLM users run windows so who knows what executable should be uploaded on HF and .exe can't be reviewed as easily as python code.

  • @isodoubIet

    @isodoubIet

    3 ай бұрын

    @@XxZeldaxXXxLinkxX They won't care, but the people developing the libraries they're relying on will care.

  • @GreatTaiwan

    @GreatTaiwan

    2 ай бұрын

    @@XxZeldaxXXxLinkxX they might not care but the one funding them will care the fact they won't have SWE by their side to help them as those cares also matters a lot

  • @aurelian891
    @aurelian8913 ай бұрын

    Contrived use case of the recursive example. In production Rust you'd use a for loop and allocate the Vec once outside and not destroy it for each iteration.

  • @XerosOfficial

    @XerosOfficial

    3 ай бұрын

    That is fair but the point is that someone in the AI space newly learning Rust will need some level of understanding these things, not to mention there possibly being other cases where the lack of tail call optimization can lead to performance issues. It's true that you still need knowhow to create performant Mojo but someone new to Mojo is less likely to fall into these obscure and minor pitfalls sprinkled throughout the journey of learning something new.

  • @Satook
    @Satook3 ай бұрын

    I’m keen to see what a really capable Mojo dev vs a really capable Rust dev can build in a fixed time window and the performance of the 2 solutions. Hell, throw a C++ and Zig dev in there too. Effort is the biggest constraint in my work.

  • @unl0ck998
    @unl0ck9983 ай бұрын

    I'm sorry but Mojician is hilariously awesome

  • @vectoralphaAI

    @vectoralphaAI

    3 ай бұрын

    Hell yeah thats what im saying. Mojicians just sounds cool at least to me. I like it better than Rustaceans and Pythonistas.

  • @richardcesar5546
    @richardcesar55463 ай бұрын

    TCO does not "unroll it into a loop" it "reuses the stack frame". When you call a non-inlined function you allocate a stack frame (by incrementing the stack pointer), this holds space for the arguments and the return value and possibly a little other bookkeeping. TCO reuses all the values in the stack frame, including the return address, and as a side effect never has to increment the stack. It can only be used in very exacting conditions (the conditions related to everything the function does must be able to be aggregated into a return value and/or argument variables, and the stack frame must be identical in size). Its real benefit is not so much performance (though it does give some) , its that it prevents a stack overflow from happening (which happens when your stack pointer overflows the preallocated stack space of your application. In this case, the vec initialization creates memory on the heap. If the next call were to rewrite the contents of that variable, the underlying vector on the heap would be leaked. Mojo is greedy in that since _stuff is never used and deleted on its "last use" there is nothing that is not stack allocated, whereas in rust it cannot. Rust can use TCO, just not with heap allocated variables. Granted, I would try an explicit drop in there after the vec init and see what happends.

  • @richardcesar5546

    @richardcesar5546

    3 ай бұрын

    The observation here would be that TCO was effective on mojo, TCO was not effective on rust, but its smart about stack sizes and had enough to not have a stack overflow, and javascript did not have TCO at all.

  • @opposite342

    @opposite342

    22 күн бұрын

    Somewhat off topic but a big thing here is that python also doesn't have TCO (at least not out of the box), so dynamic programming is better to do in mojo than python for essentially only some keywords added.

  • @remrevo3944
    @remrevo39443 ай бұрын

    45:51 Doing SIMD manually in rust really isn't fun, but often you really don't need to do it anyway. If you structure your code correctly (i.e. have it have the right alignments and sizes and stuff), LLVM is *really* good at optimizing code for your platform. So instead of doing SIMD manually you structure your code *as if* you want to use SIMD and let the compiler do the rest.

  • @CramBL

    @CramBL

    3 ай бұрын

    And writing SIMD friendly code in Rust is super easy, e.g. chunks/chunks_exact.

  • @ofekshochat9920
    @ofekshochat99203 ай бұрын

    About vec![0; 42], it actually memsets the first 42 elements. So it allocates and sets, so it might allocate on the first push. With capacity only allocates, so as long as you push less than capacity, you're guaranteed to not allocate.

  • @OnFireByte
    @OnFireByte3 ай бұрын

    Rust definitely has TCO in that example, no way that it can create like 1 billion stack without hitting stack limit if they didn’t unroll it into loop

  • @Satook
    @Satook3 ай бұрын

    I like Mojician. Also, not surprising that Chris Lattner is iterating on the intermediate language design from LLVM seeing as he was a core (founding?) dev on it for a long time.

  • @vectoralphaAI

    @vectoralphaAI

    3 ай бұрын

    Didnt Chris invent/ created the Swift programming language?

  • @rinrin4711
    @rinrin47113 ай бұрын

    There is an additional difference between semaphore and mutex. Semaphore can be controlled by anyone who got access to it, while mutex can only be released by the thread that locked it. Mutex is more of a monitor rather than semaphore.

  • @whekin
    @whekin3 ай бұрын

    I just checked the article and the benchmarks got actually updated

  • @ErikBongers
    @ErikBongers3 ай бұрын

    I don't like that "owned" thing in Mojo, because the caller of the function may not be aware that a copy of the string is being made! That's why I like that in Rust you have to explicitly use foo clone() at the caller site, making it immediately obvious to who reads that code for the first time what is going on.

  • @XxZeldaxXXxLinkxX

    @XxZeldaxXXxLinkxX

    3 ай бұрын

    They said that it transfers ownership but reverts to clone under the hood if you try to modify it, unless you use the caret in the function call, then it strictly transfers ownership (at least that's what I understood). So it seems like a quality-of-life thing, and it's up for debate for implicit vs explicit is better here

  • @XerosOfficial

    @XerosOfficial

    3 ай бұрын

    @@XxZeldaxXXxLinkxX this

  • @melodyogonna

    @melodyogonna

    3 ай бұрын

    It's always explicit when a copy isn't being made because you have to use the transfer (^) operator to move ownership. It is less clear when a copy is made vs when it's an immutable borrow, you'd have to check the original function definition

  • @orestdubay6508
    @orestdubay65083 ай бұрын

    I work with AI, I use both Python and Rust. I don't know Mojo (yet). This debate irritated me quite a bit - good debate, but I mildly disagree 🙂 We don't use python for its speed! Python is good language to "configure" frameworks like Keras, Torch, Tensorflow or Scikit - that are implemented in c++. Rust is a great replacement for that c++, not for python. Will Mojo be that c++ replacement? I have doubts. Can you trust a language rooted in Python-like prototyping to write hardcore numerical libraries? I would need some more convincing. When somebody says that it is 50% faster than Rust, that does not elicit trust - it just creates hype. On the other hand, to replace Python, Mojo would need to have a library support comparable to python - why would you use it otherwise? Again - we don't use python for its speed... Funny enough - Rust's speed or safety may as well not be the main reason to use it. I have started to rewrite some of my python code to Rust not to gain speed, but mainly for the excellent type-system and secondary for its ability to compile to wasm.

  • @adrianbergesenfedaque8016

    @adrianbergesenfedaque8016

    Ай бұрын

    agree on the type system, and add traits and pattern matching for me (I know, Python >3.11 has it too, but it feels like it was an afterthought). I like Rust approach to writing software more simply because of these language design choices (plus testing and examples). In addition, I get amazing speed and memory safety, which I welcome.

  • @CielMC
    @CielMC3 ай бұрын

    1. you can just quote the command `hyperfine "node src/index.js"` 2. The point about dropping I think is because the drop for the Vec is put after the recursive call, so it's preventing the TCO?

  • @metaltyphoon
    @metaltyphoon3 ай бұрын

    For those that don’t know C# can compile to native code just like Go. In terms of speed C#, Java and Go are very similar for many tasks. For pure synthetic benchmarks it’s C# > Go > Java. You have greater memory control in C# than the competitors.

  • @gcxs

    @gcxs

    3 ай бұрын

    C# made colored functions

  • @peanutcelery

    @peanutcelery

    3 ай бұрын

    Umm no. Is not that optimized compared to Go. You can not use a lot of libraries due to libraries not being AOT compatible. Like bro, I love C# but I’m not afraid of using another language.

  • @metaltyphoon

    @metaltyphoon

    3 ай бұрын

    @@peanutcelery neither am I but C# can give u full memory control while Go just can’t. As time passes many libs will support AOT. Minimal APIs already do it in .NET 8

  • @d1namis

    @d1namis

    3 ай бұрын

    C# is a failure as a language, it's barely floats and only because business adaption and Microsoft sales department. You are kinda right about Go > Java if you ignore most things that makes Java interesting. And comparison about GO and Java is also kinda level of understanding before junior. Did you know that you actually can tweak both GO and Java for your exact application. Discord team actually wrote a guide about optimizing GO garbage collection, and you also have CGO to go beyond that. The idea to think about C# like something more than just Office work and games for Windows is something that only C# developers can have.

  • @metaltyphoon

    @metaltyphoon

    3 ай бұрын

    @@d1namis Your take about C# is very misleading. I don't know which type of field you work on but C# is everywhere but in ML. Yes that's right, everywhere. There are many places in the world where there is more demand for C# devs than Java. That should tell you something. Go, while nimble , "easy" and easy to grok, it has major problems specially around FFI and FIPS compliance. It would be a fool's mistake to write any high performance application that needs interops with libs already written in C or C++ with Go. C# was made with this in mind ! JNI is a terrible mess, which Java added JNA. C#'s threading model is just like Rust which aligns with C. Have you ever written software that needs to be audited by the goverment? Well guess which lang is easiest to pass? All because MS has first party support for SO many thing that makes the std in Go laughable. You rarelly need third part libs. I'm telling you this as I have writen software which run on Air Force One. Web Dev is not all Dev. Anyone who thinks C# is "dying" is living under a rock for the past decade. I'm also aware of C#'s shortcommings too

  • @thomasmathews4592
    @thomasmathews45923 ай бұрын

    If you compile targeting a native CPU typically rust will auto-generate SIMD code for you, which you can see on compiler explorer with quite simple code. It becomes more fiddly if you want something that is more platform independent, or if you have dynamic input sizes which always mean you get a couple of items left at the end of the array, the remainder from array_size / sims_block_size, then you need to write painful hand-cranked stuff, but if you know what platform you are running on and compile for it you get most of the benefit without writing specialist code, just as MOJO does.

  • @remrevo3944
    @remrevo39443 ай бұрын

    31:17 The reason why rust uses drop flags is actually really interesting: If a value gets dropped conditionally, (for example when one branch has a drop(value) and another does not) rust still has to keep track of whether the value was dropped or not at the end of the scope. The solution to remove this overhead would be, to rather than keeping track of whether the value was dropped, to always drop the value at the earliest point of time that it is not used. So in the other branch that does *not* call drop(value) (As a partially deinitialized can't be used anyway, that would be totally doable.) The big downside would be that that would make the deinitialization of values slightly non-deterministic, which is also the reason this method was decided against. (Though sometimes I think it would have been nice if rust had just decided to just do the deinitialization always as soon as possible, but it might also have some other considerations like code size.)

  • @nii-san5485

    @nii-san5485

    3 ай бұрын

    you would also have to special case guard patterns (e.g. MutexGuard)

  • @remrevo3944

    @remrevo3944

    3 ай бұрын

    @@nii-san5485 Not necessarily. As long as you don't drop/move the guard (for example by giving away ownership to another function), there is not really any reason for rust to drop the guard early. But *if* you really want to always just insert drops into the program as soon as a value is no longer used (I would call that non-lexical drops), you could just have an explicit drop at the end of the guard to make sure that you explicitly use it and it doesn't get dropped early. If you want to know more about this whole topic, there is a blog article on faultlore called "Destroy All Values: Designing Deinitialization in Programming Languages" going in depth on this topic.

  • @nii-san5485

    @nii-san5485

    3 ай бұрын

    @remrevo3944 I never considered how conditional drops worked under the hood thx for linking that article. But I'm saying with your suggestion of dropping after the last use, a guard would instantly be dropped always unless an explicit drop was added or they're special cased to drop at end of lexical scope.

  • @nii-san5485

    @nii-san5485

    3 ай бұрын

    Which you did consider now that I'm reading back. I still like "lexical" drop as it's a bit more intuitive , e.g. the value just "goes out of scope" which most programmers will already have a feel for 😁

  • @foxwhite25
    @foxwhite253 ай бұрын

    That TCO example they did is so bad, Rust do have the issue of not being able to do TCO, but that example just does it, because stuff is never used or black_boxed, to demonstrate this you will need a function that counts sheep down starting at n, prints a message for each sheep counted, and returns the total number of sheep that have been counted. This function does not TCO, because the addition operation is performed after the recursive call and the result being used, so the compiler does not automatically optimize it.

  • @Reydriel
    @Reydriel3 ай бұрын

    What happened to the earlier upload? Was watching it partway, refreshed the page, and suddenly it got privated lol (it was about HTNX)

  • @liuch2000
    @liuch20002 ай бұрын

    Will not use a programming language requires my email address to install

  • @alelondon23
    @alelondon232 ай бұрын

    If the argument for migrating from python adding 15% learning to get 100x performance would be an irresistible value proposition, then everyone would be writing Nim or Julia

  • @mishaerementchouk

    @mishaerementchouk

    Ай бұрын

    Yeah, I’m quite lost in this Mojo vs Rust discussion. Which usecases we’re talking about, which developers? Say, we take the claim that Mojo has the hardware level of performance seriously. Should BLAS and TensorFlow be reimplemented in Mojo? In that case, I don’t think the familiarity would be a strong selling point. If it’s on the Python side of things, then the most of runtime is spent inside libraries anyway, so what kind of performance gain we are talking about here: instead of 7154 seconds, it will take 7127 (if we are generous)?

  • @Gregorius421
    @Gregorius4213 ай бұрын

    At 17:25 did the article confuse move semantics and copy semantics? Rust moves by default to avoid copying the string. In case of copy the original foo would be available for dbg!(foo) and there wouldn't be a compiler error. Primagen should point this out.

  • @draakisback
    @draakisback3 ай бұрын

    This article is pure marketing. These guys should have just taken the L and walked away. They make a lot of false arguments and they cherry picked that final benchmark, with something that was completely simulated. In the mojo example, the compiler actually just calls the main function and then returns because no work is being done in the program. The same happens in the rust variant when you use the vector macro instead of the with capacity call. Rust can use the same back end as mojo and it also can be optimized for SIMD. This idea that somehow a Python dev is going to have zero friction learning mojo and also get the better performance that rust is absurd. With the straight up false statements that they made in this article, I'm not going to believe anything that these people write in the future.

  • @isodoubIet

    @isodoubIet

    3 ай бұрын

    that level of skill issue when running a benchmark is indicative of one of two things, 1. incompetence when evaluating the performance of optimized code or 2. dishonesty Both make me incredibly skeptical they have the capacity to deliver on their claims.

  • @zelllers
    @zelllers3 ай бұрын

    He always reminds me of one of the voice actors in elder scrolls. Especially when he speaks the way he does in the first second of this video

  • @maniacZesci
    @maniacZesci3 ай бұрын

    Why do you harp about Arc over and over, it is a way for multiple threads/tasks to safely share same recourse, it is not specific to Rust, other languages have that too.

  • @XxZeldaxXXxLinkxX

    @XxZeldaxXXxLinkxX

    3 ай бұрын

    It's because it's a tool that effectively steamrolls over the borrow checker. Yeah there's legitimate uses, but you can just use it as a "fuck it just take the damn variable". Using it introduces overhead and performance reduction

  • @maniacZesci

    @maniacZesci

    3 ай бұрын

    @@XxZeldaxXXxLinkxX You use it when you need to use it, if you want to access same recourse from multiple threads/tasks. Yes you can misuse it but in most technologies there are tools that can be misused. My comment reflected constant Primeagen's harping about Arc like that is Rust's way of doing most of data flow/access, which is not.

  • @XxZeldaxXXxLinkxX

    @XxZeldaxXXxLinkxX

    3 ай бұрын

    @@maniacZesci he's not harping on Rust, he's harping on the the people that do that (as a crutch ) . Like harping on the people that use "as any" in typescript. Just memeing pretty much

  • @maniacZesci

    @maniacZesci

    3 ай бұрын

    @@XxZeldaxXXxLinkxX fair enough I might have missed that, not a big fan of reaction videos so I don't follow his channel closely.

  • @ClaudioBrogliato
    @ClaudioBrogliato3 ай бұрын

    To be honest though, Ruby also comes with a JIT compiler (a recent new feature) so it can be sped up if needed.

  • @kenneth_romero
    @kenneth_romero3 ай бұрын

    27:00 i believe odin does that with their built vector classes and matrices. able to get the power out of simd but have it implicitly built in. Zig is also getting them as well, but Odin is technically 1.0 already.

  • @akaidi1270
    @akaidi12702 ай бұрын

    10:35 actually got me there prime, damn, ''ruby is not skill issue, it's just slow'' Phew xD, really messed up with my head, got me in the first half, had to recheck the video for the graph xD

  • @billybest5276
    @billybest52763 ай бұрын

    I almost ended up with a Arc mutex hashmap but you saved me from that mistake a few days ago lol.

  • @michaellatta
    @michaellatta3 ай бұрын

    It sounds like mojo copy on write is like swift. What rust offers is contiguous memory layouts which gains from cache hits.

  • @PRIMARYATIAS

    @PRIMARYATIAS

    3 ай бұрын

    IIRC Matlab and Octave (its open source counterpart) also do copy on write.

  • @michaellatta

    @michaellatta

    3 ай бұрын

    @@PRIMARYATIAS Copy on write is a convenience to the programmer that has runtime performance implications. It is not a bad tradeoff for many uses, but does impact ultimate performance.

  • @user-hk3ej4hk7m
    @user-hk3ej4hk7m3 ай бұрын

    To prove the TCO example, you could write a for loop that allocates the vector the same number of times. I mean if the idea of TCO and TCE is to make recursive algorithms work as iterative then this should be a fair example of the advantages of having that optimization. My understanding is that since stack variables are eagerly destructed, every time you stop using a variable, the stack pointer decrements so when you get at the end of the funcion, your next stack starts where the old one was. This improves locality and you can work exclusively in cache, making the mojo version significantly faster, your playing with registers at that point.

  • @tychoides
    @tychoides3 ай бұрын

    I was skeptical of Mojo, but when I looked to the example in this video I said to myself: "they are mixing Rust and Python, ok lets them cook". I downloaded mojo and play with it a bit. Looks good. I am doing some test. The python library integration is ok but not perfect. The killer feature over Rust is not speed but the mojo REPL! I need a REPL for developing my data analysis code. That is the thing all those lispers were saying all this time. And Mojo gives me an REPL for dev and static compilation. That is huge win. Julia didn't have that and struggled with the JIT lag for more than 10 years due to bad language design. Now I am waiting if the guys from Modular deliver and release the language to the public (with a proper foundation) and they don't mess up the syntax too much. I am starting to feel the hype, and it worries me.

  • @rt1517
    @rt15173 ай бұрын

    28:57 What he might have omitted there is that RAII does not necessarily means heap allocation. Basically, if you are not using new, the memory for the object will be allocated in the stack. And allocating in the stack is one instruction, freeing is one instruction, no matter how many objects are allocated in the function. So this is far better than GC (if you forget that C# allows to put structs in the stack). On the other hand, yeah, malloc/free can feel slower than a GC in many cases.

  • @justanothercomment416
    @justanothercomment4163 ай бұрын

    He's saying tail call optimization isn't possible because of the scope level deferred memory managment. Meaning for that specific use tail call optimization isn't possible for Rust. Whereas for Mojo, given they are not deferring scope cleanup at the end of each call (end of scope), it easily facilitates tail call optimization. In other words Mojo will provide tail call optimization for generalized use whereas Rust only provides it in subsets where scope does not require memory allocation clean up until scope destruction. This is also why the "drop" didn't fix the issue because it's not changing the deferred scope destruction. In Rust, "drop" flags the allocation for scope destruction. Effectively changing nothing.

  • @OnFireByte

    @OnFireByte

    3 ай бұрын

    1. Rust is definitely doing TCO in the example he show since there's no way that the program will create 1 billion stack without hitting the limit. 2. I don't see any reason that allocating heap memory in recursive function will make TCO impossible in rust, since converting TCO-able function into loop is literally just declaring argument as normal variable, then put the function body into the while loop, and compiler can just put the clean-up step before jumping to next loop.

  • @justanothercomment416

    @justanothercomment416

    3 ай бұрын

    @@OnFireByte I agree, I would very much like to better understand what's going on here. It could be the author is falsely attributing TCO failure to some underlying semantic implementation detail within Rust. I guess a review of the generated code is the only way to know for sure what's actually going on there.

  • @GrantGryczan

    @GrantGryczan

    3 ай бұрын

    3. Drop flags only show up when you have a drop that can't be statically determined (e.g. if a variable is only dropped when a runtime condition is true). There are not going to be any drop flags compiled into that code, and explicitly adding an unconditional drop before the recursive call _should_ cause the implicit drop at the end of scope to be omitted.

  • @justanothercomment416

    @justanothercomment416

    3 ай бұрын

    @@GrantGryczan My understanding is the drop always adds a bit flag and nothing more. The drop is then evaluated at end of scope. Which means in this case, the drop remains regardless of the drop flag. Resulting in no change as the drop is already deferred from scope destruction. In other words, the drop is saying I want you to do what you're already planning on doing.

  • @GrantGryczan

    @GrantGryczan

    3 ай бұрын

    @@justanothercomment416 As I said in my last comment, that is incorrect; the drop flag is only needed in very few cases. Take a look at the official nomicon documentation on drop flags. It explains this well and is very short and easy to understand.

  • @marble_wraith
    @marble_wraith3 ай бұрын

    I'm all about the idiot-matic...i write code, come back 6 months later and think: "which idiot wrote this?...oh"

  • @pluto8404

    @pluto8404

    3 ай бұрын

    then you rewrite it better, comeback in 6 months and think "which idiot wrote this" and rewrite it how you had it the first time.

  • @georgesanchez8051

    @georgesanchez8051

    3 ай бұрын

    That’s me after 48 hours

  • @sovereignberserk1346

    @sovereignberserk1346

    3 ай бұрын

    Haha okay found my peer group in here 🤣👌

  • @az-qf9ht
    @az-qf9ht3 ай бұрын

    Bro I’m never forgetting what a mutex or semaphore is after that godly explanation. 13:00

  • @az-qf9ht

    @az-qf9ht

    3 ай бұрын

    @@harikrishnanb7273 13:00 :D

  • @ikixox5054

    @ikixox5054

    3 ай бұрын

    That “explain it to me like i am 4 because i am too dumb to be 5” had me crying

  • @r.pizzamonkey7379
    @r.pizzamonkey73793 ай бұрын

    So the problem with tail-call optimization in this instance is that they added an extra semicolon, that's it. That's a feature that's in C as well, tail-call optimization only happens when you're returning the final expression. Also, the reason Vector::new() is faster is the allocation gets optimized away.

  • @michaellatta
    @michaellatta3 ай бұрын

    GC requires indirect access. Direct allocation/deallocation can cause fragmentation. Rust tends to have larger continuous struts than copy on write memory management. Explicit memory management can run in far smaller memory usage.

  • @AM-yk5yd
    @AM-yk5yd3 ай бұрын

    ~40:00 Maybe rust in current iteration of the compiler in release mode can detect that vec is not used and DCE it out of existence. More proper test would be to do something with vec. Honestly, at that time it'd be necessary to look at assembler code. (I'm not fan of their copies, it seems it can create a lot of headaches, if some things will be copied deeply, some not, but hopefully they thought of it and there will be no auto_ptr 2.0 but with every copyable type)

  • @patronspatron7681
    @patronspatron76813 ай бұрын

    Modular pinned Primeagen who then passed Modular by reference.

  • @BaardFigur
    @BaardFigur2 ай бұрын

    How would mojo work e.g. with scope_guards? Are there any ways to extend the lifetime to the end of the scope?

  • @gandalfgrey91
    @gandalfgrey9117 күн бұрын

    Looking back at that blogpost, there is a very incorrect usage of commas throughout, and it dips its toes into Oxford commas seemingly at random. Doesn’t Modular have anyone to proofread?

  • @danilogomes-br
    @danilogomes-br3 ай бұрын

    What is the Rust book shown at the beginning of the video? 5:03

  • @digama0

    @digama0

    3 ай бұрын

    zero to production in rust, it's mentioned on the screen

  • @lptimey
    @lptimey3 ай бұрын

    42:52 my guess is that, because they delete objects as soon as they arent in use, and becausethe vec‘s never in use, that they never alocate until u write code that uses it.

  • @hatonafox5170
    @hatonafox517029 күн бұрын

    I'd love to see Prime react to an interview Chris Lattner did on how Mojo works and what's happening behind the scenes. I don't think most people realize that Mojo is simply using Python as the syntactical glue for a completely different set of backend processes. To use a car analogy, if Python is a Toyota Camry it's syntax is the paint job and decaling on the outside of the car. Mojo is a Formula 1 car that uses the same paint color as Python's Toyota Camry but it also has cool racing stripes and decals for the superset features/syntax. When I hear people talk about Mojo it's as if they think the language is still a Python Toyota Camry with some aftermarket mods to make it faster which leads them to think it can't possibly be as fast as something like Rust which to them is racecar. Nope Mojo is an actual Formula 1 car with a familiar Python paint job. That's about all they have in common. Which clears up why there's so much appeal. Mojo is basically telling Python programmers you just need to learn a few new concepts and some additional syntax and you'll be able to drive this Formula 1 car. What's even better is that it'll feel almost as easy to drive as your Toyota Camry which you can still drive whenever you want.

  • @maxlee3838
    @maxlee38382 ай бұрын

    There’s a “tailcall” crate which adds an annotation (derived trait) to functions.

  • @Mn16Cr45
    @Mn16Cr453 ай бұрын

    CNC maschines are programmed with C here in germany. Depending on cpu and how mojo works their will be a bright future for it. But for I dont know how this will be on a cnc maschine cuz they need execute one after the other and not parallel thats how a cnc works

  • @ViaConDias
    @ViaConDias3 ай бұрын

    The Vec with capacity allocates the vector whereas the new Vec is removed by the compiler because it's never used. I do not know Rust but from a general compiler viewpoint, this would be logical. Rust might even "zero" out the memory allocated to the Vec of capacity. Mojo seems to just wait to allocate until a value is pushed to the vector meaning it never allocates any memory for the Vector in the given example.

  • @darsoma95
    @darsoma953 ай бұрын

    I have no idea what's going on in this video, but I find it fascinating

  • @tswdev
    @tswdev3 ай бұрын

    If Mojo can have Pydantic data structs with validation, HTTP libs for serving and posting, database connectors and a kafka connector or something in addition to the AI stuff on the standard library, it could potentially be THE lang for AI powered web

  • @vectoralphaAI
    @vectoralphaAI3 ай бұрын

    I honestly LOVE that Mojo programmers are called Mojicians. It just sounds cool. A programming magician. Honestly sounds better than Rustaceans and Pythonistas.

  • @brod515
    @brod5153 ай бұрын

    the point prime made at @19:24 about the ownership being orthogonal to the type is actually quite good. I wish rust did this the other way round. It seems the fell into a trap trying to make references similar to C++ references. the could required you to say things like `ref` and `owned/copy/clone`. and also remove the idea of implicit copies and require you to always .clone() something.

  • @clementdato6328
    @clementdato63283 ай бұрын

    One thing i don’t like about rust is it is not predictable what the compiler does in many cases in terms of optimization. I frequently find some weird not-optimized-away reason here or there in the user forum. But i guess this improves over-time.

  • @batboy49
    @batboy493 ай бұрын

    I hate bash too...btw anybody know how to get scp to recognize * wildcards in zsh? For some reason that will not work for me. Is there some plugin I need?

  • @brod515
    @brod5153 ай бұрын

    They've actually changed the article on part for Tail Call Optimization. it might be worth rereading that part.

  • @Amasglobulaires
    @Amasglobulaires3 ай бұрын

    Has he already covered with the community the Julia Language ?

  • @BakaNeko15
    @BakaNeko153 ай бұрын

    wait i am watching him after like 6 months, why does he have a lettuce on his head??

  • @desertfish74

    @desertfish74

    3 ай бұрын

    Valid concern.

  • @baltoky11

    @baltoky11

    3 ай бұрын

    lost a bet

  • @andrewdunbar828

    @andrewdunbar828

    3 ай бұрын

    freedom

  • @BakaNeko15

    @BakaNeko15

    3 ай бұрын

    @@andrewdunbar828 fair enough

  • @oleksiistri8429
    @oleksiistri84293 ай бұрын

    respect to mojo for using "fn" instead of "def"

  • @serena_m_

    @serena_m_

    3 ай бұрын

    It actually has both-def remains the same as regular Python, while fn has new Mojo semantics & optimizations

  • @the_mastermage

    @the_mastermage

    3 ай бұрын

    @@serena_m_ which in my humble opinion is worse. Sorry but making two different ways to write functions will be so confusing. They also have two types of objects, the standard class and structs. I think this is messy and will make things more difficult, for people coming from python.

  • @isodoubIet

    @isodoubIet

    3 ай бұрын

    fn is bad because it's the nth element of the f sequence. Are we paying by the character now?

  • @Seedzification
    @Seedzification3 ай бұрын

    great we needed one more language, can't wait for the next one

  • @pashadia
    @pashadia3 ай бұрын

    Hey, I wonder if you could have a proc macro in Rust that would get passed a standard function and would transform all `i32`, `f64`, etc into their SIMD counterparts behind the scenes

  • @rudyperalta5592

    @rudyperalta5592

    3 ай бұрын

    There's stuff for better SIMD ergonomics in nightly Rust actually. It's std::simd

  • @Charles-um6qo

    @Charles-um6qo

    3 ай бұрын

    This is literally not necessary. LLVM can unroll / vectorize for loops. In Rust, explicit SIMD actually just maps to LLVM intrinsics anyway. You can manually unroll loops yourself / with macros, which might help. Mojo is not adding anything, auto vectorization is nothing new and explicit SIMD is outside the scope of what a scientist would want to be doing.

  • @skeleton_craftGaming
    @skeleton_craftGaming3 ай бұрын

    Correct C++ is Garbage collected, kinda... There's these things called smart pointers that use the constructor/destructor paradigm to automatically delete on out scope

  • @OnFireByte

    @OnFireByte

    3 ай бұрын

    Rust also has that, it’s RC/ARC. Definitely not a traditional GC like GC tier language (mark and sweep that need to stop the world) but yeah you could say that

  • @skeleton_craftGaming

    @skeleton_craftGaming

    3 ай бұрын

    @@OnFireByte I'm not say rust doesn't have it I'm just saying calling C++ a manual memory language is wrong if you follow best practices (which is to not use raw pointers unless you don't transfer ownership).

  • @AM-yk5yd

    @AM-yk5yd

    3 ай бұрын

    That's a lie. If C++ had GC it would be possible to make equivalent of python's class GraphNode: linked: List["GraphNode"] Impossible in C++, You need to come up manually with explicit strategy on who owns what in a graph and clear up memory. * You can't use unique_ptr's because many nodes can link the same node. * You can't use shared_ptr because graphs have cycles which means if you have A->BC and pass A and A goes out of scope, B and C survive * It's impossible to use weak_ref because somebody need to have non-weak ref. So you need to manually make graph class to handle ownership because C++ has no GC. You don't need to do anything of that in GC. "My program doesn't leak memory kinda" doesn't count. when it ~kinda does.

  • @skeleton_craftGaming

    @skeleton_craftGaming

    3 ай бұрын

    @@AM-yk5yd no it's not they're called smart pointers.

  • @OnFireByte

    @OnFireByte

    3 ай бұрын

    @@AM-yk5yd it’s just how you define GC, many people consider reference counting as GC because they define GC as just a system that automatically and safely deallocating memory at runtime, but yeah RC isn’t GC if you wanna say that GC need be able to deallocate every data thats doesn’t get referenced by root node (tracing GC). It’s just definition anyway

  • @olazawho
    @olazawho3 ай бұрын

    I was curious and did some reading, because I was curious how Mojo can claim pass-by-reference as a default, and also better semantics. Maybe folks in the know could clear things up for me? I see that Mojo currently has implementations of neither explicit lifetimes, nor enforcement around taking immutable references when a mutable reference is still alive. I'm also unclear on how Mojo ending lifetimes at point of last call makes reference lifetime semantics easier to reference about. Does that just mean that every reference is an RC by default? Also, doesn't the eventual implementation of mutable borrow enforcement have the potential to introduce a lot of complexity into this system? And, if passing by `owned` sometimes references, and sometimes moves, doesn't that also force the programmer to understand how the Mojo compiler works to identify performance bottlenecks, and gets an extra level of complication when mutable borrows are enforced? It must be that Mojo will eventually *actually* move the value if it's mutated and has other references alive, so that introduces a third branch of implicit behavior that you might have to track. At least, presumably, package authors will. It's logically impossible that Mojo can be faster than Rust (or even faster than Go above a certain level of borrowing complexity?), have less implicit copying (or much more use of reference counting), and have simpler ownership semantics, right? Unless they've found some new proofs around ownership that unlock fundamentally different approaches to resolving ownership. I just want to make sure I'm not missing something/too stupid to understand it before I form a strong opinion about how much this cart is being put before the horse. EDIT: I see now that Mojo doesn't support *returning* references, because that's obviously what causes the need for explicit lifetimes. That removes the needs for RCs, surely reasoning about lifetimes becomes *more* difficult, and not less, when ownership doesn't necessarily last until the end of a scope?

  • @olazawho

    @olazawho

    3 ай бұрын

    I guess if the idea is that Mojo's explicit purpose is supporting ML, and isn't worried about being natively integration into larger application stacks, a lot of those things might not be issues. Seeing as you're unlikely to have to worry about mutable references and having large numbers of RCs when you're mostly doing matrix operations

  • @retereum

    @retereum

    3 ай бұрын

    Wait, isn't that reference behavior is a standard python thing? I see no difference here, in python unless you start explicitly modify variable, it's passed as ref(for example, pop() and append() on the list if you don't assign smth to that reference).

  • @olazawho

    @olazawho

    3 ай бұрын

    @@retereum mojo isn't garbage collected, it's borrow checked, like rust is. So you have to know (whether through the user manually tracking, or through proofs built into the compiler) that the underlying memory you're pointing to is still safe to access

  • @isodoubIet

    @isodoubIet

    3 ай бұрын

    @@olazawho Unless they want to be in a similar situation as python where the code eventually gets rewritten as C++, they'll have to write a bunch of business logic around it. It better be at least as good as Rust for that. That business logic probably won't be the perf bottleneck but being able to reason about ownership and lifetimes is still important. I don't think reasoning about lifetimes is any more difficult than with scoping, but you do lose out on functionality (can't stick cleanup logic in an ad-hoc destructor, can't tie it to locks, etc). IMO it's not worth it.

  • @mateusvmv
    @mateusvmv3 ай бұрын

    You use hot-reloading to solve compile times in the debug builds It's the best of both worlds

  • @kaishang6406
    @kaishang64063 ай бұрын

    recently doing vhdl for a course. A simple clock have a complie time of 10 minutes.

  • @mrdatapsycho
    @mrdatapsycho3 ай бұрын

    One more Important thing to realise, If you know python and have learned Rust, you are more close to learn Mojo. Because Mojo also introducing features from Rust like Ownership and Borrowing etc. Adding such features will have a skill issue impact on Python developer interested in learning Mojo. Because al-least you need to learn those concept before using them.

  • @leonardeuler4
    @leonardeuler43 ай бұрын

    Where can we see the source code of the mojo compiler?

  • @JoseDaviCorreiaPassos

    @JoseDaviCorreiaPassos

    3 ай бұрын

    Not open source yet

  • @d3stinYwOw

    @d3stinYwOw

    3 ай бұрын

    @@JoseDaviCorreiaPassos Probably won't be for a long time :)

  • @victorcadillogutierrez7282

    @victorcadillogutierrez7282

    3 ай бұрын

    That's the neat part, you can't :).

  • @kinomonogatari

    @kinomonogatari

    3 ай бұрын

    In the best case scenario it would follow the Swift timeline where the language is open sourced after around 1.5 years after first appearing. In the worst case scenario you end up with a situation similar to MATLAB. Although Chris has said Modular is aiming to open source the stdlib by Q1 2024 (December 2023 Modular event)

  • @andrewdunbar828

    @andrewdunbar828

    3 ай бұрын

    @@kinomonogatariBest case should be better than the Swift case since that was down to Apple whereas Chris can now release it as soon as he thinks it's the right time.

  • @MetroidChild
    @MetroidChild3 ай бұрын

    "Future proof for 50 years" sounds like a dumb prediction, that includes a ramp up and ramp down of usage like we've seen with C, and by the time those 50 years are reached (if that even happens) a new and better language will have been developed. There is nothing lost by learning the language, especially if you already use python, but it's a hyperbolic statement.

  • @KimberlyWilliamsch
    @KimberlyWilliamsch3 ай бұрын

    the compatibility of linux distros is only limited to ubuntu, as a debian user I haven't experienced what it's like to program using mojo.

  • @yash1152
    @yash11523 ай бұрын

    32:20 TCO i know, but what is TCE tail call elimination?

  • @larryrowe
    @larryrowe2 ай бұрын

    As a mainframe coder I used Java/several other packages/ assemble/cobal (note it can talk to rust code)/etc./c/c++(overloaded, virtual memory problems) ... Java to talk to Rust, Go to manage HTML NONSENSE, ETC., ...

  • @MikkoRantalainen
    @MikkoRantalainen3 ай бұрын

    47:30 I think "Would you switch to typescript if introducing this new syntax would allow it to run 100x faster than javascript?" would be even closer analogy. And even I would stop writing vanilla JavaScript if TypeScript were actually faster.

  • @rogerdinhelm4671
    @rogerdinhelm46712 ай бұрын

    If we start to account for skill issues, then Java can be as fast as Rust/C++ or even faster (after warmup), because having enough skill you can write garbage-free code and make mnual memory allocations/deallocations. And the part that can make it faster is JIT optimizations, which can be done in current specific usecase, like look-unwinding or operation reordering, which C++ or Rust simply cannot do, because they don't know how the code they produce will be used every time you run a program.

  • @HyperCodec
    @HyperCodec3 ай бұрын

    Lol his hair is slightly transparent bc of the green screen

  • @larryrowe
    @larryrowe2 ай бұрын

    For games use their custom interfaces along with Mojo/Java/etc.

  • @jagagemo8141
    @jagagemo81413 ай бұрын

    I make "normal" software for work, cant remember the lasttime speed was a big issue. Don't get me wrong, we need it to not be a hot trash fire, but that's it.

  • @brod515
    @brod5153 ай бұрын

    @22:13 I don't quite get that point they made. if every object has an identity then there must be some indirection happening there. if every object has an identity then that means there is a lookup to get any object ( aside from direct memory location ). If the lookup is insignificant then the argument is fine. but if you always have to first lookup the objects id then get it's location in memory that's obviously a penalty that you don't have to pay with Pinning in rust. The idea of pining is that data can be at 0x01 and it moves to 0x9. any one who assumed it was at 0x01 needs to know that it's now at 0x9. pinning allows you to not have allocate dynamic memory all the time.

  • @thugspeedman7869
    @thugspeedman78693 ай бұрын

    I have no clue what any of this means but I can't stop watching.

  • @EwokPanda
    @EwokPanda3 ай бұрын

    I think Primeagen misread that as "Explain that like I'm 5 years into a Computer Science program"

  • @MikkoRantalainen
    @MikkoRantalainen3 ай бұрын

    From the official Mojo manual: "Mojo uses a third approach called “ownership” that relies on a collection of rules that programmers must follow when passing values. The rules ensure there is only one “owner” for each chunk of memory at a time, and that the memory is deallocated accordingly. In this way, Mojo automatically allocates and deallocates heap memory for you, but it does so in a way that’s deterministic and safe from errors such as use-after-free, double-free and memory leaks. Plus, it does so with a very low performance overhead." So it's much closer to Rust than Java or J# or JS.

  • @larryrowe
    @larryrowe2 ай бұрын

    I don't why you can't use Go and Rust together as appropriate, java/Rust, HTML/etc. Mojo ... (mainframe does do assemble), Mojo , heavy data management: Rust ,Java for human interactive code to the other systems.

  • @real1cytv
    @real1cytv3 ай бұрын

    So, as far as I understand it, Rust doesn't implement TCO when you are allocation because it isn't really "safe" and can lead to unintended behavior. The article"The Story of Tail Call Optimizations in Rust " has a little bit on this, though it's quite old. The reason Vec::new doesn't do this is because Vec::new only lazily allocates, therefore TCO can be applied, but Vec::with_capacity DOES do heap allocation and according to the rust devs if they did TCO there, this might lead to undefined behavior. Thoguht the speedup when using vec! is wierd, since vec! is just using Vec::with_capacity and fill under the hood... Maybe some optimizations?

  • @alomac8976
    @alomac89763 ай бұрын

    You should write an LSP in C++

  • @DrewryPope
    @DrewryPope3 ай бұрын

    c# is actually faster than go out-of-the-box and can drop down to unsafe code and manual memory to go even faster.

  • @stevenhe3462
    @stevenhe34623 ай бұрын

    About TCO, Rust _does not guarantee_ it, though LLVM _may_ do it. Rust does not have stable TCO because it drops the stack variables at the end of the scope, after the recursion call.

  • @AM-yk5yd

    @AM-yk5yd

    3 ай бұрын

    Rust is working on proper TCO(they chose terrible keyword 'become' (which replaced even more terrible keyword `be`)) where possible, see issue #112788 in rust-lang. Which is pending because LLVM is buggy again.

  • @stevenhe3462

    @stevenhe3462

    3 ай бұрын

    @@AM-yk5ydThis would probably be one of the features that are pending forever.

  • @HasSF
    @HasSF2 ай бұрын

    If there is one person qualified to make a new language it's chris lattner. These discussions are better approached with a very fresh perspective + very open mind. that's the only way to properly evaluate them and truly understand the essence behind the point that the other side is making

  • @Gregorius421
    @Gregorius4213 ай бұрын

    17:58 `copy` would be a more intuitive (less confusing) choice for the `owned` keyword.

  • @XxZeldaxXXxLinkxX

    @XxZeldaxXXxLinkxX

    3 ай бұрын

    Get owned

  • @LtdJorge
    @LtdJorge3 ай бұрын

    Something must have gone wrong. All vec![v; n] does is call vec::from_elem(v, n), which calls Vec::with_capacity(n) and on the returned Vec, Vec::extend_with(n, v). How can just calling Vec::with_capacity(n) be worse than that?

  • @MoncefNaji
    @MoncefNaji3 ай бұрын

    Java and C# run on a runtime which is faster than running on an interpreter but not as fast as it could be by running straight on the metal, like compiled languages

Келесі