Accelerated Training by Amplifying Slow Gradients

Grokfast: Accelerated Grokking by Amplifying Slow Gradients
arxiv.org/abs/2405.20233
github.com/ironjr/grokfast/tr...
Support my learning journey either by clicking the Join button above, becoming a Patreon member, or a one-time Venmo!
/ tunadorable
account.venmo.com/u/tunadorable
Discuss this stuff with other Tunadorks on Discord
/ discord
All my other links
linktr.ee/tunadorable

Пікірлер: 120

  • @ironjr
    @ironjr10 күн бұрын

    Author here! Happy to have this wonderful introduction. Huge thanks for reviewing our work! Your explanation looks way better than mine 😂.

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    Happy you liked it! I just saw your tweet as well. Feel free to get in contact if you'd like to hop on a video call for a short interview to post on the channel

  • @breeanabellflower

    @breeanabellflower

    10 күн бұрын

    Hi @ironjr - I loved the paper. @tunadorable did a great job with explaining it too. It feels like this is a significant find. I really want to see if applying grokfast adaptively as suggested could just be a generally good idea. The examples suggest that this could work on pretty broad problems. I may end up testing this myself later tonight, but I want to get your thoughts Jaerin!

  • @modernsolutions6631

    @modernsolutions6631

    10 күн бұрын

    @ironjr Low pass filters can be easily implemented using a simple ODE and don't require keeping any history. When designing filters one has flexibility with regard to filter order and so on. Is there a reason why don't implement a construction for arbitrary filters? Do you think an adaptive filtering approach (in particular: the further the training accuracy lacks behind the validation accuracy the more high frequencies are filtered) could be interesting? Also can you distinguish the effect of increasing weight decay and lowpass filtering the gradients of the overfitted network? Does that tell us something interesting?

  • @RuslanLagashkin

    @RuslanLagashkin

    10 күн бұрын

    Your paper is impressive and well written. Can't wait to test the approach in action.

  • @lukeskywalker7029

    @lukeskywalker7029

    9 күн бұрын

    @ironjr thx for this amazing work! Cant wait to try replicating this on a bit bigger LLMs. Would you say the suggested ranges for hyperparams should stay the same?

  • @dr.mikeybee
    @dr.mikeybee11 күн бұрын

    What a smart and simple idea! Just adjust down big errors and adjust up small errors. This keeps the weights from jumping around.

  • @HuyTruong-bd4hb

    @HuyTruong-bd4hb

    10 күн бұрын

    it's smart but not simple lol it took many researchers in many years of deep learning experiments with the grokking phenomenon to come up with the idea 😅

  • @minecraftermad

    @minecraftermad

    10 күн бұрын

    what exactly do you mean by this interpretation? would something as simple as: range = max(gradients) - min(gradients); adjustedGradient = sign(gradients[i]) * ((abs(gradients[i] / range) - 0.5) ^ 2 + 0.5) * range; weight -= 0.01 * adjustedGradient; work? or is there a more complex idea behind the comment?

  • @minecraftermad

    @minecraftermad

    10 күн бұрын

    seems i've misunderstood what you meant, it's the error variance in the time dimension. though i think my thing can be rotated, but then it will probably take more memory. i'll try it next though.

  • @amortalbeing

    @amortalbeing

    10 күн бұрын

    the marvelous thing about it is the perceived simplicity, yet no body could come up with this until these gentlemen had this bright idea! everything is hard until someone solves it and then the masses wonder why this seemingly simple thing! has not yet been solved by them before also sometimes, the building blocks had to be layed down for others to come to a specific conclusion/understanding, this is one of those examples as well. still lovely work both in previous papers and this one concerning grokking!

  • @thetrueanimefreak6679
    @thetrueanimefreak667911 күн бұрын

    Love your videos man keep it up, cant believe I found you on accident never looked back

  • @alexanderbrown-dg3sy
    @alexanderbrown-dg3sy10 күн бұрын

    My guyyyy. You’re too real for this 😂. Yea ways around the memory tibits, ternary LM’s are the future and few other tricks makes this less of a problem. Data quality is just as important, like extremely important, hybrid data is the future(“Grokked Transformers are Implicit Reasoners” empirically prove this), people called me looney but I was like raw data is way too oblique, but we must figure out a derivative of MLE that’s optimized for model collapse, completely solvable….but I stand on it. Paper of the year 😂. People are sleeping. Deeper models are actually more susceptible to grokking, but the compute needed is an issue, this may be the solution. A super cheap POWERFUL hobbling as the kids say. Research like this will allow smaller labs/startups to compete with the big guys. Compositional generalization is within reach.

  • @oncedidactic

    @oncedidactic

    10 күн бұрын

    With mechinterp upsides eh?

  • @GabrielVeda
    @GabrielVeda10 күн бұрын

    Very impressed by this overview.

  • @GermanTopGameTV
    @GermanTopGameTV8 күн бұрын

    If someone asked my what "Grokking" was before I saw this video I would have assumed it was some teenage slang meaning something like "wearing a piece of Salad on your head and attracting all the girls" but I seem to not only fall out of style with the youth but also with the recent trends in tech.. I get old..

  • @highviewbarbell

    @highviewbarbell

    5 күн бұрын

    "grok" comes from the sci fi book Stranger in strange Land

  • @Embassy_of_Jupiter
    @Embassy_of_Jupiter10 күн бұрын

    I guess grokking makes sense. It's way easier to memorize than to understand and once you memorized everything, you can fit all the pieces together much easier.

  • 10 сағат бұрын

    Are you talking about humans?

  • @graham8316
    @graham831610 күн бұрын

    Replacing the high frequency gradients with different types of noise would be interesting

  • @lukeskywalker7029

    @lukeskywalker7029

    8 күн бұрын

    What do you think could be the benefit of re-adding white, pink or some other noise spectrum? Dont really get the intuition...

  • @graham8316

    @graham8316

    8 күн бұрын

    @@lukeskywalker7029 like do we just need slow gradient with jiggle, or do the specific fast gradients from the training batch help. If we just need slow gradient and jiggle maybe we can learn much faster than we have been

  • @lukeskywalker7029

    @lukeskywalker7029

    8 күн бұрын

    @@graham8316 can you give me a python or even better triton implementation for that? Then ill test it out ;)

  • 9 сағат бұрын

    Perhaps use something inspired by dropout?

  • @DerAlbi
    @DerAlbi10 күн бұрын

    Viewing this as an electrical engineer with some knowledge about signal processing, it seems like the next intuitive step is to implement a Biquad-IIR-Filter with per parameter(-group)-tunable frequency cut-off for a given architecture. It takes more memory again, as an IIR-filter has more state again, but the frequency response is much more controllable than an EMA. For example implementing a Low-Shelf-Filter is exactly what is needed here. This paper basically went from FIR-Filtering (a MA is a very stupid FIR-filter choice) to an IIR-Filter structure which is more memory efficient while preserving the frequency behavior of a MA. I suggest to make the IIR-Filter more complex again and go from first order to 2nd order or even concatenating 2 Biquads to get a 4th order response. I also wonder if this applies to gradient-solvers in general or if this a NN-specific optimization. For example, does it help a least-suqare-function-fitter too, or can it help a solver for the sparse matrices of FEM-problems? I wish you would have explained the "Groking" more, as I basically still dont know what it is, and the paper's explanation is way too targeted at an audience with a certain knowledge base and googling did not help much either.

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    here’s a tiktok explaining it www.tiktok.com/t/ZPRE8S1eS/ and i can’t understand what you said but ill also mention the author so that they see this @ironjr

  • @marcfruchtman9473

    @marcfruchtman9473

    10 күн бұрын

    That tiktok is a pretty good explanation! I think the one important point that might be missed is that "Grok/Grokking" comes from Heinlein's 1961 book, Stranger in a Strange Land, where it is used as a Martian term for "to understand intuitively or deep understanding". Yea, I didn't get it until I understood that little bit. Basically once the model has trained far beyond the normal amount, it can "sometimes" reach this point of grokking where it gets an emergent deep understanding. The sad part tho, is that this phenomena, at least so far, is not applicable to general learning, but rather very specific algorithms... like the Modulus algorithm, at least from what I have read so far. It is one of the reasons why I tend to ignore "grokking" titles in videos, but I wanted to see what this paper was about! It seems pretty good, tho, without a general application, I am not sure when it can be used since over training is likely to be very expensive, and if there's not a usable "grokking" to come about,. then you are just wasting cycles.

  • @Tunadorable

    @Tunadorable

    9 күн бұрын

    I've left similar explanations in other comments: it's nowhere near certain yet but given the fact that grokking can be induced in non-synthetic datasets (such as MNIST & IMDB sentiment classification displayed in this paper), the current speculation is that in these synthetic datasets it's just easier to see grokking since they only cover a *single* skill/algorithm, where as in something like language modeling there are *many* skills/algorithms to be learned. If true, that would mean that in LLM loss curves the grokking effect becomes invisible since you've got so many tiny sub-skills continuously grokking that they get lossed in the average and give you a seemingly smooth loss curve with only a slight delay, and that this phenomenon would explain why we have smooth loss curves and yet sudden emergence of individual skills on benchmarks

  • @marcfruchtman9473

    @marcfruchtman9473

    9 күн бұрын

    @@Tunadorable Thanks for your insight. It would be interesting to test this process on a highly focused Domain Specific Language Model, where it might be possible to see Grokking for specific training (something bigger than IMDb but still small enough to see those effects).

  • @DerAlbi

    @DerAlbi

    9 күн бұрын

    Thank you guys so much for the productive comments! That TikTok was all I needed. Super cool.

  • @maxkuc2470
    @maxkuc24706 күн бұрын

    Thank you so much for highlighting tools like this! Keep it up!

  • @joe_limon
    @joe_limon10 күн бұрын

    If training gets efficient enough. I can foresee mini training cycles between user prompts could allow models to internalize their interactions. And allow them to actively become specialized toward helping a user.

  • @lukeskywalker7029

    @lukeskywalker7029

    8 күн бұрын

    already venturing into that with a few use cases 7 clients!

  • @buybuydandavis
    @buybuydandavis9 күн бұрын

    Like update momentum, but taking an average instead of using geometric decay?

  • @q44444q
    @q44444q9 күн бұрын

    How is this any different from stochastic average gradient (SAG) methods?

  • @amortalbeing
    @amortalbeing10 күн бұрын

    Thanks a lot that was very interesting!

  • @leastofyourconcerns4615
    @leastofyourconcerns46156 күн бұрын

    Do you maintain some kind of list of topics and papers you've been working through by any chance?

  • @Tunadorable

    @Tunadorable

    5 күн бұрын

    I wish I had been doing so from the beginning. Eventually I plan to open-source all the pdf highlights & notes in my notes app but we're not quite there yet. For now you can check out the file `papers_kept.csv` in this repo to see a list of all the papers that I plan to read which gets updated every week (first week was last week) github.com/evintunador/arxiv-summaries-workflow

  • @AndreInfanteInc
    @AndreInfanteInc4 күн бұрын

    Wait, I'm confused. If you're just holding an exponential moving average in memory and adding it to the gradient, how is that different from, like, regular inertia?

  • @wwkk4964
    @wwkk496411 күн бұрын

    Awesome! Fits well with the recent ANU researchers who demonstrated that you can induce grokking in arbitrary systems by overparameterizing and the notion is model agnostic. Video here kzread.info/dash/bejne/h3aeybWfh8SYpNo.htmlsi=T6cjE8f0UENxLqSr

  • @DanFrederiksen
    @DanFrederiksen6 күн бұрын

    Isn't this essentially like momentum way back in the day?

  • @zacharris6299
    @zacharris629911 күн бұрын

    I wonder whether the EMA approach would allow for second order training methods with small batch sizes to work well. Would be awesome to try L-BFGS with EMA and small batches in micro GPT sandbox!

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    working on an update to the templateGPT thing right now to put it more in line with karpathy’s latest video (meaning it’ll be more efficient & also capable of training ~100m parameter models on a gpu cluster) and i’m gonna include this grokfast-EMA thing as an optional add-on that you can toggle in the settings

  • @zacharris6299

    @zacharris6299

    10 күн бұрын

    @@Tunadorable Nice. I've been playing with it on a 3090 and can train up to ~50m somewhat reasonably on that with it. It definitely doesn't support L-BFGS as is though and some of the unusual models like future former apparently store some tensors on the CPU which leads to a failure so I have to turn GPU off

  • @Tunadorable

    @Tunadorable

    8 күн бұрын

    yes i’ve not yet got FutureFormer working, hoping to refactor it after i finish the current templateGPT edits. excited to hear someone is actually using these😁

  • @litical3719
    @litical37194 күн бұрын

    Thankyou for breakdown. As a non-machine learning person this was really interesting and seems ground breaking.

  • @PeterKornezos
    @PeterKornezos4 күн бұрын

    Is there an implementation for tensorflow? I am new to machine learning and I am familiar only with tensorflow.

  • @Tunadorable

    @Tunadorable

    4 күн бұрын

    You might try searching on GitHub. However the code I linked to is simple enough that you could copy & paste it into chatGPT or Claude 3.5 and ask it to translate from pytorch to tensor flow and I'd bet the latter especially could get it right first try

  • @PeterKornezos

    @PeterKornezos

    3 күн бұрын

    @@Tunadorable I am going to try that !

  • @danielcogzell4965
    @danielcogzell49657 күн бұрын

    Please keep us up to date with any applications of grokfast and whether or not it improved performance for you 😊

  • @timothywcrane
    @timothywcrane11 күн бұрын

    Blah Blah Blah..... LSTM FAST-GROKKING? Hell Yeah! Great info. I get excited looking back at things "b4 transformers" simply because they cannot and are not the end all - be all in pretraining. I especially love LSTM (xLSTM and this lately piqued my interest) variants surrounding music and action (task,scene, whatever - just please, not another stock algo) prediction/suggestion (Is your film "too" predictable?).

  • @yakmage8085
    @yakmage80858 күн бұрын

    Nice review of the paper, sub unlocked

  • @wadyn95
    @wadyn958 күн бұрын

    Does it applies to other tasks/datasets?

  • @Tunadorable

    @Tunadorable

    8 күн бұрын

    To my memory they tested three tasks/datasets in this paper of reasonably wide variety (synthetic math, image classification, and language sentiment analysis) but as with all science, we won't know if / how well it works with others until somebody tests it out and publishes

  • @815TypeSirius
    @815TypeSirius10 күн бұрын

    At this point we arent going ahead without this so anything that can improve scaling and training is super important to replace humans with intelligent life.

  • @verigumetin4291

    @verigumetin4291

    10 күн бұрын

    lol

  • @815TypeSirius

    @815TypeSirius

    10 күн бұрын

    @verigumetin4291 caps like "ai scawy," but humans have committed all known atrocities, and ai has done zero.

  • @Tunadorable

    @Tunadorable

    8 күн бұрын

    i beg to differ, today the walmart self checkout camera AI caught me trying to counteract inflation and i’d count that as a crime against humanity

  • @815TypeSirius

    @815TypeSirius

    8 күн бұрын

    @Tunadorable The fake ai apocalypse is going to be hilarious. When Trump is elected again he spends so much money on fake ai to throw people in jail. Its like Hitler with less charisma.

  • @jawadmansoor6064
    @jawadmansoor60648 күн бұрын

    can it be applied to (how) to CNNs and RNNs?

  • @Tunadorable

    @Tunadorable

    8 күн бұрын

    The code can certainly be dropped in & used on a CNN or RNN but they did not test it in this paper, so results may vary. In theory there's no reason why it wouldn't also work there just as well

  • @beschterrowley3749
    @beschterrowley37497 күн бұрын

    what is a basic mental model to interpret all this, or even just grok itself, for a newbie

  • @Tunadorable

    @Tunadorable

    7 күн бұрын

    intuititve metaphor explanation of grokking here: www.tiktok.com/t/ZPREjCckt/

  • @beschterrowley3749

    @beschterrowley3749

    7 күн бұрын

    I don't got no tiktoks this is KZread

  • @tvwithtiffani
    @tvwithtiffani9 күн бұрын

    This is probably a silly question but can someone explain to me how the technique of Grokking and the ai hardware company known as Groq both came up with the same phonetically sounding names? Is it as simple as someone within Groq capitalizing on the word Grok(king)? Just very curious. not looking for conspiracies, just probable reasons.

  • @Tunadorable

    @Tunadorable

    9 күн бұрын

    not sure how Groq chose their name, but the original term “grokking” was coined by a 1961 sci fi author where he defined it as “to understand deeply” and the model training behavior that was named after that sci fi author’s term was discovered by researchers at Apple

  • @tvwithtiffani

    @tvwithtiffani

    9 күн бұрын

    @@Tunadorable Oh nice. I thought it would be too much to ask what grokking actually meant. Thank you, I appreciate this.

  • @MooseEngineer
    @MooseEngineer11 күн бұрын

    "I SKIPPED RIGHT BY IT" You seem to skip a lot of things, helluva smart young fella.

  • @wanfuse
    @wanfuse10 күн бұрын

    where do you publish papers anyway?

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    i have never published a paper, but 99.9% of the ones i read in these videos come from arxiv.org

  • @wanfuse

    @wanfuse

    10 күн бұрын

    @@Tunadorable any idea on what happens to intellectual property for papers published there?

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    @@wanfuse no clue, but I imagine it'd be the same as any other pre-print website

  • @ckq
    @ckq2 күн бұрын

    EWA seems kinda obvious, no? That's the best estimate for a signal that is signal + noise. Would've thought that's already included in Adam and the like, with all the complicated momentum terms .

  • @ckq

    @ckq

    2 күн бұрын

    I say this cause I've been using EWA and similar variants to model signals in my free time for 4+ years now.

  • @wege8409
    @wege84096 күн бұрын

    Will try out this optimizer when I get to work, curious to see how it does vs. classic Adam

  • @AP-dc1ks
    @AP-dc1ks10 күн бұрын

    So its just momentum.

  • @chadwick3593

    @chadwick3593

    10 күн бұрын

    That's what I was thinking. I guess now we have a new way to understand momentum.

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    so i’m not an optimizer person but they clarified this toward the end. to the best of my memory they said that rather than being momentum on the parameter movement its momentum on the gradient itself, which is somehow slightly different. i could be getting fooled on this tho, like i said optimizers are not my thing

  • @chadwick3593

    @chadwick3593

    10 күн бұрын

    @@Tunadorable Momentum always refers to adjusting the gradient based on 2nd order estimates, usually with something like an exponential moving average of the gradient, like what this paper does. AdamW, which I think is currently the most popular optimizer, does that as well. AdamW divides out the variance instead of low-pass filtering out frequencies, but the effect should be the same: slow-moving gradients get amplified, and fast-moving gradients get dampened. AdamW also includes weight decay, which this paper finds useful, and it corrects for the statistical bias that comes from averaging gradients from random minibatches, which this paper doesn't do. (Look up Bessel's Correction on Wikipedia. Same idea.) I do find the signal theory framing for understanding "fast-moving gradients" useful, and I do find it useful to see the relationship between fast/slow moving gradients and grokking. It's unclear to me if the training algorithm is useful, but the framing certainly is.

  • @zacharris6299

    @zacharris6299

    10 күн бұрын

    @@chadwick3593 Looks to me like if there is a significant difference, it's the adjustable window size and application of a low pass filter to several previous steps at once, whereas AdamW has to rely on two moments to represent the entire past.

  • @garazaadsf6921

    @garazaadsf6921

    10 күн бұрын

    momentum is all you need

  • @marcfruchtman9473
    @marcfruchtman947310 күн бұрын

    Apologies to the authors, as I realize that my comment might sound a little harsh. I am sure the research is well founded. But, isn't Grokking a super niche emergent effect that doesn't really apply generally to LLM, but only under very specific circumstances, such as very specific math algorithms? I guess to me, this type of research really needs a proof that it can be applied more generally, before it will have significant utility. When I first started to see "grokking", it was quite the buzz word, but the more I learn about it the less excited I am. I can see the allure, just keep training far beyond normal and "eventually" your model learns something. But it sounds to me like this is a carrot that will never be able to applied without a very specific set of data and specific training requirements. Until they can find some more general application, I see very little utility in spending that much more extra time to over train, just to find out that the Grokking isn't going to work unless it is some esoteric math function that no one will actually use for a real application. (But, if I have it wrong, then I apologize). Edit: I guess since it does show some efficacy in the IMDb sentiment classification. So, perhaps I was too hasty in my earlier comment. I think I had trouble deciphering the graph for the IMDb, with the overlapping colors for raw and EMA.

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    perfectly fine critique. my impression right now is that grokking is visible on those single synthetic math tasks specifically because they are *single* tasks that the model is being trained on, in contrast to LLMs which have to do *many* tasks. speculation in the field right now is that grokking may in fact happen at the level of individual tasks within an LLM, but there are so many tasks in there that everything ends up averaging out and we get only a marginal if not invisible delay in test accuracy. after all, in this paper they also trained an MNIST classifier and an IMDB sentiment analysis model which, while by no means are they as complicated as an LLM, certainly are examples of real data with multiple tasks that still experience grokking and can have that grokking corrected / sped up by this method

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    i meant to add that if this speculation about LLMs is in fact true, then it would mean that grokking is an explanation for sudden skill emergence despite smooth loss curves

  • @marcfruchtman9473

    @marcfruchtman9473

    10 күн бұрын

    @@Tunadorable Ah, valid points, tho memory might be a issue with larger data sets. I recall that selective State Space models seems to be more efficient with memory, but, I haven't seen anything about Grokking for those. If you happen to come by anything... it might be worth a review.

  • @Tunadorable

    @Tunadorable

    9 күн бұрын

    to the best of my memory that's exactly the process that this paper was describing kzread.info/dash/bejne/gYKom9VwdJDYepc.html

  • @wege8409

    @wege8409

    6 күн бұрын

    We've gotta do what we can, if it's possible to see improvements gained from a novel approach to a niche problem then it might be possible to amplify that solution in some way to where it generalizes better.

  • @Biedropegaz
    @Biedropegaz10 күн бұрын

    Nice and thx for the recording, but if You need a banana :-o I will send u one.

  • @user-qw1rx1dq6n
    @user-qw1rx1dq6n10 күн бұрын

    Isn’t this just momentum?

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    so i’m not an optimizer person but they clarified this toward the end. to the best of my memory they said that rather than being momentum on the parameter movement its momentum on the gradient itself, which is somehow slightly different. i could be getting fooled on this tho, like i said optimizers are not my thing

  • @zacharris6299

    @zacharris6299

    10 күн бұрын

    Looks to me like if there is a significant difference, it's the adjustable window size and application of a low pass filter to several previous steps at once, whereas AdamW has to rely on two moments to represent the entire past.

  • @screwnacorn

    @screwnacorn

    4 күн бұрын

    Alg 1 is just some weird form of dual averaging, Alg 2 is basically just momentum... nothing new here

  • @sarthakmishra1483
    @sarthakmishra148311 күн бұрын

    This may be as big as NAGD , ADAM etc.

  • @Daniel-Six
    @Daniel-Six10 күн бұрын

    Anyone find it interesting that grokking only works reliably on synthetic data?

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    As this paper mentions it's been shown on MNIST and IMDB sentiment analysis, and to the best of my memory the first paper that explored those two datasets found a reliable phase-shift between good training, grokking, and divergence when they did a grid search over multiple different hyper parameters (weight decay and learning rate to the best of my memory). IF it's happening in LLMs the thought is that they have so many tasks (these synthetic datasets are each only one "task") that each individual grok occurrence ends up averaging out and it becomes invisible in the loss curves, but it does constitute a possible explanation for sudden skill emergence despite smooth loss curves

  • @TheOlderSoldier
    @TheOlderSoldier4 күн бұрын

    The whole concept of high vs low frequency gradients makes me think of training wheels on a bicycle. Initially, you need the harsh feedback of the training wheels until eventually you are able to maintain balance with finer internalized corrections.

  • @rey82rey82
    @rey82rey824 күн бұрын

    Grokking unlocked

  • @joe_limon
    @joe_limon10 күн бұрын

    Current training sequences are incredibly inefficient. Think about how little data we as humans consume to get to our intelligence. Many many orders of magnitude less data than these models.

  • @Tunadorable

    @Tunadorable

    10 күн бұрын

    I've heard this argument before but I've gotta disagree because it's always framed in the lens of one single modality. Sure if you compare the amount of language an LLM has trained on to the amount of language a 5 year old has heard there's no comparison, the 5 year old looks more efficient. But in reality the 5 year old was also taking in extremely high resolution vision, stereo audio, touch, taste, smell, and proprioception every millisecond of being awake since being born. I saw a tweet awhile ago, I think by Yann Lecunn but I could be misremembering, that tried to estimate how much multimodal data a 5 year old has taken in and the low estimate completely blew out of the water our current 15T token LLMs. That plus the human brain like all animal brains is pre-wired with certain phenomena already innately known, such as a fear of snakes, whereas neural networks are randomly initialized. That being said I do agree with you that current training strategies are likely massively inefficient compared to what we'll have in the future

  • @joe_limon

    @joe_limon

    10 күн бұрын

    @@Tunadorable the vast majority of data humans intake is immediately lost due to attention mechanisms, and then further pruned during sleep cycles. Contrast the amount of visual information taken in by your eyes every second, and then think back to the last time you were in a car. How much of that trip do you remember? How many red lights vs green lights?

  • @gasun1274

    @gasun1274

    9 күн бұрын

    Humans are purely analog machines. We don't have limited word lengths or an instruction set abstraction between training models and the actual silicon doing the math.

  • @keanurust4946
    @keanurust494611 күн бұрын

    oi

  • @TheDoomerBlox
    @TheDoomerBlox3 күн бұрын

    2:28 - thanks to it being a logarithmic scale, you can - for once - actually say that it is an order of magnitude faster. because it is (actually) an order of magnitude faster, which is pretty significant, yes

  • @erickmarin6147
    @erickmarin61479 күн бұрын

    Lmao so grokking is enhanced by some kindof low pass filter

  • @biomedicalaibasics
    @biomedicalaibasics6 күн бұрын

    Fun fact: If you double tap a comment, the comment gets liked 👍

  • @sikunowlol
    @sikunowlol11 күн бұрын

    oi