Sholto Douglas & Trenton Bricken - How to Build & Understand GPT-7's Mind

Ғылым және технология

Had so much fun chatting with my good friends Trenton Bricken and Sholto Douglas on the podcast.
No way to summarize it, except:
This is the best context dump out there on how LLMs are trained, what capabilities they're likely to soon have, and what exactly is going on inside them.
You would be shocked how much of what I know about this field, I've learned just from talking with them.
To the extent that you've enjoyed my other AI interviews, now you know why.
There's a transcript with links to all the papers the boys were throwing down - may help you follow along.
Website & Transcript: www.dwarkeshpatel.com/p/sholt...
Spotify: open.spotify.com/episode/2dtD...
Apple Podcasts: podcasts.apple.com/us/podcast...
Trenton Bricken's twitter: / trentonbricken
Sholto Douglas's twitter: / _sholtodouglas
Timestamps:
(00:00:00) - Long contexts
(00:17:04) - Intelligence is just associations
(00:33:27) - Intelligence explosion & great researchers
(01:07:44) - Superposition & secret communication
(01:23:26) - Agents & true reasoning
(01:35:32) - How Sholto & Trenton got into AI research
(02:08:08) - Are feature spaces the wrong way to think about intelligence?
(02:22:04) - Will interp actually work on superhuman models
(02:45:57) - Sholto's technical challenge for the audience
(03:04:49) - Rapid fire

Пікірлер: 335

  • @Flyingblackswan
    @Flyingblackswan28 күн бұрын

    This podcast isn't just a podcast, it's a valuable insight to how brilliant people think. This could've easily been like a $1,000 course. I really appreciate the guests for their time and for your ability to conduct an excellent interview.

  • @danielbrockman7402

    @danielbrockman7402

    28 күн бұрын

    this is so true

  • @aoeu256

    @aoeu256

    23 күн бұрын

    yep! however, i was thinking why don't we have a model of "emotional speech" yet where each word can be spoken with different emotions, speeds, voices. maybe the problem is the speed of our keyboards.

  • @JoblessGuyReading

    @JoblessGuyReading

    22 күн бұрын

    @aoeu256 yeah I think Hume ai is actually trying to do that , saw a few of their demos !

  • @MrEmbrance

    @MrEmbrance

    22 күн бұрын

    Jesus, stop

  • @Randy.Bobandy

    @Randy.Bobandy

    11 күн бұрын

    You're got big problems if you'd pay $1000 to watch this.

  • @aidanmclaughlin5279
    @aidanmclaughlin527929 күн бұрын

    This is, by far, the densest and most enjoyable interview I've ever heard in this domain. Please get these guys back on.

  • @sup3a
    @sup3a29 күн бұрын

    I don't understand how this podcast isn't bigger. It's like Lex but better. Keep doing what you do, you rock dude

  • @ryzikx

    @ryzikx

    28 күн бұрын

    because lex is not "better" he is more popular. he is for a more general audience

  • @egalanos

    @egalanos

    28 күн бұрын

    I think posting lots of little clips would be hurting viewing stats for the algorithm. I really wish there was a separate clips channel.

  • @KibberShuriq

    @KibberShuriq

    25 күн бұрын

    I find this particular episode extremely fascinating, but I also feel like I personally know maybe one other person I could recommend watching it to, and even they would probably understand only about half of it. Lex, on the other hand, is usually very legible to an average (or slightly above average) native English speaker. Also, Dwarkesh never asks his guests what love is, and we all know that's pretty much a requirement for popularity.

  • @HarpreetSingh-xg2zm

    @HarpreetSingh-xg2zm

    25 күн бұрын

    @@egalanoslittle clips are more likely to funnel new viewers to this channel. Channel is too small to try and branch clips out

  • @TomBouthillet

    @TomBouthillet

    23 күн бұрын

    Lex is an insufferable douche nozzle.

  • @akhil090579
    @akhil090579Ай бұрын

    you should do more of these double interviews I think it worked really well and moves in interesting directions, good work!

  • @Sirbikingviking
    @Sirbikingviking29 күн бұрын

    Wow this is one of the most dense and interesting podcasts I have ever seen, and my Spotify account has over 700 episodes listened to. Great job Dwarkesh, thank you so much for what you are doing.

  • @hummuswithpitta
    @hummuswithpitta24 күн бұрын

    This is everything Lex's original "AI Podcast" should have been. You are so god damn well researched and the interviewees know it and relish in your questions, which just gives us so much extra sauce. Hope you just focus on the general AI universe and not bother diluting the podcast like Lex has (zero shade to Lex here btw). These two are honestly fascinating to listen to. Crazy how short-lived their career in the industry is for the knowledge they have.

  • @andrew-729

    @andrew-729

    20 күн бұрын

    Nah, cast shade on lex. He never actually pushes back, and when he does it is super light. Basically they know it is a free for all PR circuit.

  • @sidequestsally

    @sidequestsally

    5 күн бұрын

    "I'm just making context queries" -Dwarkesh probably

  • @JohnSimerlink2014
    @JohnSimerlink2014Ай бұрын

    When Jim from the Office and Justin Timberlake teach you AGI

  • @DwarkeshPatel

    @DwarkeshPatel

    Ай бұрын

    lmao

  • @eliz3225

    @eliz3225

    3 күн бұрын

    justin timberlake talks like mark zuckerberg

  • @JazevoAudiosurf
    @JazevoAudiosurf16 күн бұрын

    "it's amazing how easily you can become world class at something because most people are not trying that hard" it's amazing how easily you can be in the top 10% because nobody is trying at all

  • @siddharth-gandhi
    @siddharth-gandhiАй бұрын

    you da mad king of podcasting. putting in dem mad hours before each. thanks for being awesome!

  • @DwarkeshPatel
    @DwarkeshPatelАй бұрын

    Hope you guys have as much fun with this as I did :)

  • @LoganJeya

    @LoganJeya

    Ай бұрын

    Would love to see you talk to Joschua Bach if you haven't already. These two have such a great intuitive understanding of the mind and how these learning systems work

  • @space_ghost2809

    @space_ghost2809

    Ай бұрын

    Try to have Ilya again.

  • @Ashish-yo8ci

    @Ashish-yo8ci

    Ай бұрын

    @@space_ghost2809 or Jan Leike 😅 They have some interesting stuff going on in the superalignment wing with scalable optimisation and weak to strong generalization.

  • @13371138

    @13371138

    Ай бұрын

    This channel is amazing in terms of the guests, depth of discussion, and production quality. I am completely baffled as to why this channel does not have 1M+ subscribers. Maybe because the channel thumbnail makes it look too casual?

  • @XTargi

    @XTargi

    Ай бұрын

    Yeah , that's super nice. Can you please consider adding some acoustic panels on the wall, because it's echoing pretty hard 😅

  • @andrewwalker8985
    @andrewwalker8985Ай бұрын

    Spectacularly interesting and entertaining interview. These guys are smart in the way that makes you feel like you’re in on it

  • @BillyBarnyarns
    @BillyBarnyarnsАй бұрын

    Dwarkesh most of this went over my head, but when I could follow along it was very enjoyable. Some of your questions were exceptional. In terms of constructive feedback, there were a few moments when you personally chose to explain a key concept for the audience, rather than allowing the experts to make the explanation. I would have preferred if you had delegated that role to the experts, as they are better suited to making a complex concept more accessible to beginners. I also felt that you spoke slightly slower than usual in this video, which I also appreciate. Makes following the conversation much easier. Finally, it is obvious that you have put in a lot of work to try and understand this space and I absolutely commend you for that. Congrats on a great video!

  • @JC-ji1hp
    @JC-ji1hp28 күн бұрын

    Hands down the best episode I’ve watched yet. That’s saying a lot for this podcast. Great work Dwarkesh!

  • @jeffreypicel7863
    @jeffreypicel786327 күн бұрын

    One of the best AI podcasts in the past year!

  • @duudleDreamz
    @duudleDreamz29 күн бұрын

    Amazing! One of the best vids/interviews out there about AI

  • @DwarkeshPatel

    @DwarkeshPatel

    29 күн бұрын

    Glad you enjoyed it!

  • @henry3435
    @henry343528 күн бұрын

    Fantastic, Dwarkesh! This episode is my favorite podcast I’ve listened to in years.

  • @Shaunmcdonogh-shaunsurfing
    @Shaunmcdonogh-shaunsurfing29 күн бұрын

    This and Ilyas interview with Jensen are up there as the most insightful interviews as to how AI works outside of simple neural nets imo

  • @jaybazad6292

    @jaybazad6292

    17 күн бұрын

    same observation, there were sections in Ilyas interview that laid out the path forward, so much spun around his words. And now this.

  • @human_shaped
    @human_shapedАй бұрын

    This is one of the best conversations so far.

  • @DentoxRaindrops
    @DentoxRaindropsАй бұрын

    Great podcast as always, please never stop posting these videos, Dwarkesh!

  • @DwarkeshPatel

    @DwarkeshPatel

    29 күн бұрын

    More to come!

  • @mikestaub
    @mikestaub29 күн бұрын

    The fact that Trenton is visibly worried about interpretability is incredibly scary to me, given how well he understands the technology.

  • @wffff2
    @wffff221 күн бұрын

    This is one of the most interesting interview I've seen on your channel. I don't know who they are and haven't googled about them yet, but they are amazing. The progress that they are describing is the most accurate interpretation of AI advancement that I've heard.

  • @puneettripathi740
    @puneettripathi74010 күн бұрын

    Hey buddy thank u for your content I have seen you for first time n i really appreciate the people u bring into conversation, your content motivates me to learn and I am indebted to you for that. 🙏🏽

  • @mikestaub
    @mikestaub29 күн бұрын

    You are doing a great job with these interviews Dwarkesh, keep it up!

  • @ussgordoncaptain
    @ussgordoncaptainАй бұрын

    This took me 8 hours to finally finish as I constantly had to read the references just to barely comprehend the conversation What kind of brain power are we dealing with

  • @starsandnightvision

    @starsandnightvision

    28 күн бұрын

    Even the humor is on another level. I was like what is so funny about this or that lol

  • @schwajj

    @schwajj

    27 күн бұрын

    Just your typical 99th percentile humans who have spent thousands of hours learning their craft in a community of others like them. Edit: or maybe 99.9%

  • @zhedrag04

    @zhedrag04

    25 күн бұрын

    ​@@schwajj More like 99.99 the average human is borderline mentally handicapped

  • @aoeu256

    @aoeu256

    23 күн бұрын

    99.99%@@schwajj

  • @raybrandt
    @raybrandt28 күн бұрын

    It's outrageous how good this podcast is.

  • @EvolHeartriseAI-qn5oi
    @EvolHeartriseAI-qn5oi9 күн бұрын

    This has got to be the best episode you've done so far. The pure concentration of knowledge in this episode far surpasses any other AI/ML podcast I've seen thus far. Please get these guys on again.

  • @alexmolyneux816
    @alexmolyneux81624 күн бұрын

    Really love you putting in the time for this. Cannot overstate how much this adds to my life as an AI practitioner

  • @vascodegraaff
    @vascodegraaff12 күн бұрын

    One of the best episodes I've listened to all year! Keep up the great work

  • @GNARGNARHEAD
    @GNARGNARHEAD29 күн бұрын

    incredible conversation, would love to see more of 'em, thanks

  • @nitap109
    @nitap109Ай бұрын

    Excellent, Dwarkesh you did well.

  • @mohammadkazemsadoughi3880
    @mohammadkazemsadoughi3880Ай бұрын

    I have watched the first 15 min and it’s amazing! Thanks for sharing these great contents. Hope to finish watching it today :)

  • @chefatchangs4837
    @chefatchangs483725 күн бұрын

    I’ve finally found THE podcast for ML. Thanks Dwarkesh. Truly amazing content.

  • @bgtyhnmju7
    @bgtyhnmju727 күн бұрын

    Pair of great guys. Some great information and insight into the industry. Thanks for the video

  • @husainzaidi
    @husainzaidi29 күн бұрын

    subtitles was a life saver to understand this podcast. But good stories!

  • @dr.mikeybee
    @dr.mikeybee29 күн бұрын

    This is a fascinating discussion. I love that associative memory is linked to fully connected NNs while semantic nearness is done in attention heads. As we traverse the existing codification of neuroscience for its functional mechanisms, we can think about the appropriate connectionist and symbolic data structures.

  • @prakadox
    @prakadox26 күн бұрын

    Great podcast! As we wrangle with these new entities LLMs , conversations like this which lay out the various challenges are super useful.

  • @pianoforte611
    @pianoforte61127 күн бұрын

    I've never watched Dwarkesh before but this was incredible. Deep insightful questions that show how much effort he puts into getting the most out of these interviews. And Trenton and Sholto are two brilliant and passionate people. 2:55:00 seems like the most important of the discussion - how can you be sure you can control a superhuman model? And the solution is to identify circuits that allow it to lie or do other malicious things. But I'm not completely sold on the proposed solution. It seems that a truly superhuman model should be able to simulate a plausible feature map that shows when it's lying - you would then ablate it and think the problem is solved. But the model is actually hiding where it's true capabilities are. It could even simulate the behavior you would expect if your ablation worked, so you can never know if it's keeping something dormant for future use.

  • @nonstandard5492

    @nonstandard5492

    26 күн бұрын

    I think you're right, but I think a bigger issue comes to head before you even get that far. Dwarkesh has tried to ask about it a few times, including in this episode, but it always gets brushed aside. Superhuman AIs will have I guess "more conceptual space", in silicon valley speak, than we do. In other words, they'll have concepts we don't and maybe even can't understand, so how the f are we supposed to look for the dangerous ones?

  • @JonathanPlasse
    @JonathanPlasse29 күн бұрын

    Thank you for this awesome episode

  • @TheMirrorslash
    @TheMirrorslash22 күн бұрын

    Please bring them back after GPT-5 has been out. This podcast is so valuable, goated.

  • @seanfuller3095
    @seanfuller309524 күн бұрын

    Best YT vid I’ve seen seen something Andrej did 2 years ago probably. Thank you

  • @euromaestro
    @euromaestro10 күн бұрын

    Really enjoyed this. Would be great to include links to papers discussing ICL and GD correspondance/relationship.

  • @maximetouze9669
    @maximetouze966928 күн бұрын

    Sick episode thank you! Easy to understand but so interesting too, bravo

  • @samlouiscohen
    @samlouiscohen29 күн бұрын

    Phenomenal interview!!

  • @JumpDiffusion
    @JumpDiffusion28 күн бұрын

    This was packed. Great stuff.

  • @marky8078
    @marky807826 күн бұрын

    I want more!! I have 2 full notebooks of notes. Great podcast.

  • @pratikdagu
    @pratikdagu17 күн бұрын

    The knowledge density in this episode is ..😮. I am not in the cs or ml field but the things these guys said makes so much sense like the superposition paper and sparsity penalty..makes it intuitive to understand.

  • @PseudoProphet
    @PseudoProphet11 күн бұрын

    LLMs are much smarter than humans already. We just don't know how to utilise them properly yet. I think in future even a 7b parameters model will be better than any human in doing all tasks.

  • @yonatan09
    @yonatan0927 күн бұрын

    Such a perfect episode. From the camera angles and beautiful people to the topics to the clarity of explanations. Wow 3 hours just flew by.

  • @TheManinBlack9054
    @TheManinBlack905428 күн бұрын

    Underrated channel

  • @daveh56821
    @daveh5682126 күн бұрын

    🎯 Key Takeaways for quick navigation: 00:52 *🗣️ Introductions and Achievements* - Introduction of Sholto Douglas and Trenton Bricken, highlighting their contributions to AI, particularly in AI alignment and mechanistic interpretability. - Sholto's role in Gemini's success despite being relatively new to the field. 02:10 *🧠 Context Lengths and AI Intelligence* - The underappreciated importance of long context lengths for AI models. - The significant impact of increasing context lengths on model intelligence, demonstrated through the ability to learn a new language and potentially play Atari games. 05:00 *🔍 In-Context Learning and Gradient Descent* - Exploring in-context learning as a form of gradient descent. - The potential for in-context learning to lead to superhuman capabilities by integrating massive amounts of information. 07:19 *🎓 Long-Horizon Tasks and AI Agents* - The connection between long context windows and AI's ability to perform long-horizon tasks. - Addressing the misconception that the inability to perform long-horizon tasks is the primary reason AI agents haven't progressed significantly. 10:42 *⏳ Understanding Long-Horizon Task Success Rates* - The importance of evaluating AI models on long-horizon tasks to gauge their economic impact and capability improvements. - Challenges in measuring success rates over tasks with varying time horizons. 13:21 *🤔 Learning in the Forward Pass* - The shift towards more learning happening in the forward pass of models. - Comparing learning processes to natural evolution and considering the implications for AI efficiency and adaptability. 17:28 *🧠📚 Reasoning vs. Raw Information Storage* - Differentiating between the storage of raw information and the process of reasoning within AI models. - How models transform input tokens into meaningful outputs through layers of processing. 24:08 *🧩 Pattern Matching and Intelligence* - The role of pattern matching and associative memories in intelligence. - How high-level associations and meta-learning contribute to AI's reasoning capabilities. 27:02 *🔍 Sherlock Holmes Analogy and Deductive Reasoning* - Discussing the deductive reasoning abilities of AI in the context of Sherlock Holmes. - The importance of context length and working memory in enabling complex reasoning and problem-solving in AI. 31:44 *🤖 Superintelligence Concerns and Associations* - Debating whether the associative nature of AI reasoning should alter concerns about superintelligence. - Considering the implications of AI's capabilities being grounded in associations and pattern recognition. 32:43 *🔄 Recursive Self-Improvement and Associations* - Discussion on the recursive self-improvement of AI focusing on enhancing association skills. - AI's potential for meta-learning, implying an ability to improve its association-making skills rapidly. 33:41 *🚀 Intelligence Explosion Theories* - Examination of intelligence explosion theories from the perspective of AI researchers. - The plausibility of an intelligence explosion driven by AI researchers being replaced by automated systems. 37:51 *🧠 AI Augmentation in Research* - How AI can augment research work by automating tasks, potentially speeding up AI research progress. - The importance of more reliable models and the potential for AI to automate significant portions of research tasks. 44:05 *🛠️ Research and Experimentation Process* - Insight into the process of AI research, emphasizing experimentation, idea validation, and understanding failures. - The significant role of intuition and experience in selecting and executing research ideas. 50:25 *📈 Scaling AI Research* - Challenges and strategies in scaling AI research, including organizational and computational aspects. - The importance of compute resources and the concept of "taste" in driving research direction. 53:18 *🤖 AI's Role in Accelerating AI Research* - Discussion on how AI could directly contribute to accelerating AI research, focusing on algorithmic progress and the production of synthetic data. - The distinction between AI augmenting researchers' capabilities and AI's output being a crucial component in model capability progress. 59:39 *🔍 Empirical Nature of ML Research* - The empirical and evolutionary nature of machine learning research, suggesting a gradual, community-driven path towards advanced AI models. - The impact of increasing participation in the field on the pace of AI advancements. 01:05:36 *📈 Future AI Jumps and Brain Scale Comparisons* - Discussion on future improvements in AI capabilities and comparisons with human brain scale. - Even with diminishing returns on compute investment, significant capability improvements are expected. 01:07:02 *🧠 Sample Efficiency and Model Size* - Examination of sample efficiency and the impact of model size on learning. - Larger models exhibit greater sample efficiency, potentially addressing data efficiency challenges. 01:10:52 *🤖 Model Interpretability and Compression* - The challenge of interpreting highly parameterized models and the role of compression in model training. - Techniques for improving model interpretability by manipulating activation spaces and reducing compression. 01:13:25 *🧩 Adaptive Compute and Chain-of-Thought* - Adaptive compute as a method for models to allocate more processing to complex problems. - Chain-of-thought reasoning as an example of adaptive compute, allowing models to "think" through problems over multiple steps. 01:23:20 *🔄 AI Agents and Communication Efficiency* - The potential for AI agents to communicate using dense representations for efficiency. - Future AI systems may utilize more human-interpretable features for internal communication. 01:25:08 *💡 Near-Term AI Firms and Agent Specialization* - Near-term AI systems likely to resemble networks of specialized, reliable agents. - Importance of human oversight in early AI systems to ensure alignment with desired outcomes. 01:29:19 *🌐 Language Evolution and AI Development* - The co-evolution of language and human cognitive capabilities as a model for AI development. - Language's role in structuring thought and its potential influence on the effectiveness of language models. 01:36:06 *🚀 Career Reflections in AI Research* - Reflections on the rapid progress and contributions to the field of AI research within a short career span. - The importance of execution, experimentation, and a proactive approach in advancing AI interpretability research. 01:38:20 *🔄 Career Agility and Headstrongness* - Discussion on the significance of being headstrong and able to pivot careers or academic focuses with agility. - The value of strong ideas, held loosely, allowing for swift directional changes when necessary. 01:43:28 *🛠️ Building a Career through Agency and Novel Contributions* - How unique contributions to AI and engineering can significantly boost visibility and career opportunities. 01:53:55 *🧠 Bridging Knowledge Across AI Subfields* - Reflecting on the synergy between computational neuroscience insights and AI model interpretability research. 01:59:48 *🚀 Leveraging Passion and Expertise Beyond Formal Channels* - The non-linear paths to significant roles within major tech organizations through passion-driven projects and external recognition. 02:09:26 *🧠 Brain vs. AI Model Features* - Exploration of how features in AI models and the brain might correlate, questioning the granularity and universality of features across both. 02:15:28 *⚙️ Feature Universality and Misalignment Concerns* - Discussion on feature universality across models and its implications for AI interpretability and alignment. 02:25:51 *🔍 Analyzing AI Models through Dictionary Learning* - Introduction to using dictionary learning for interpreting and understanding AI models by identifying and analyzing features. 02:32:18 *🤖 Feature Splitting and Scalability in AI Interpretability* - Examination of feature splitting as a critical aspect of AI model scalability and interpretability. 02:40:57 *🔍 Depth-First Feature Exploration* - Discussion on a methodical approach to identify highly specific features, like bioweapons, in AI models by incrementally increasing the dimensionality. 02:45:58 *🎛️ Mixtral of Experts and Feature Organization* - Examination of how features and experts within models are organized, questioning the intuitiveness and specialization of these features. 02:58:52 *💡 High-Level Associations in AI Interpretability* - Delving into high-level feature associations found in AI models and their potential parallels with human psychology. 03:03:05 *🤔 The Ethics and Potential Risks of Too Much Control* - Raises concerns about the ethical implications and potential risks of having too much control over AI models through precise interpretability techniques. 03:10:29 *🍦 AI Enjoyment of Task Prediction* - Speculative discussion on whether AI models might "enjoy" predictable sequences, comparing it to human preferences for predictability. Made with HARPA AI

  • @kabir09999
    @kabir0999914 күн бұрын

    Jeez!! Why am I having a man crush on both of them

  • @riley.matthews
    @riley.matthewsАй бұрын

    Great as always.

  • @Macorelppa
    @MacorelppaАй бұрын

    Dwarkesh Patel transformed me from an A.I. hater to an A.I. lover! 🙌

  • @avefreetimehaver5154

    @avefreetimehaver5154

    Ай бұрын

    One look at that kingly face bro

  • @bestboy007

    @bestboy007

    Ай бұрын

    why? ugay?

  • @avefreetimehaver5154

    @avefreetimehaver5154

    Ай бұрын

    @@bestboy007 oops forgot to say no homo

  • @andybaldman

    @andybaldman

    Ай бұрын

    I'm so sorry to hear that.

  • @bestboy007

    @bestboy007

    Ай бұрын

    ?@@andybaldman

  • @georgewashington7251
    @georgewashington725128 күн бұрын

    Really god interview. A lot of good data that will take a while to fully appreciate.

  • @midgetsanchez
    @midgetsanchez26 күн бұрын

    Both Sholto & Trenton strike me as the type of folks to leave and start their own AI companies in the near future ;)

  • @TheOddy80
    @TheOddy8028 күн бұрын

    The level of this conversation was truly refreshing after listening to the same surface level theories and explanations over and over since GPT-3.

  • @TimeLordRaps
    @TimeLordRaps28 күн бұрын

    Dwarkesh's request for research at the head of an episode is a genius idea.

  • @badrraitabcas
    @badrraitabcas23 күн бұрын

    Great episode. Would be cool if you brought people who have been playing around with the Mamba model. Its memory costs scale linearly unlike transformers and it performs as well on language modeling. It's pretty fast during inference as well as training too. The quadratic costs of transformers might become too big of a bottleneck for the industry in the near future.

  • @dr.arslanshaukat7106
    @dr.arslanshaukat710624 күн бұрын

    Sholto is damn good looking dude for sure.

  • @Michaelhajster
    @Michaelhajster29 күн бұрын

    Huge value!🙌🏽

  • @JonathanJoergensen
    @JonathanJoergensen29 күн бұрын

    This was great! thanks

  • @effestop
    @effestop28 күн бұрын

    Bro, amazing dense (and sparse) interview. Thx.

  • @simoneromeo5998
    @simoneromeo599828 күн бұрын

    Great podcast. One hour into it, it seems to me that we are missing a point when it comes to intelligence explosion. The speakers are guessing what it would look like only when software improvements allow for it, but we should keep into consideration hardware improvements too. AI will help us to design better chips, discover superconductors, new energy sources, new manufacturing techniques. This will give AI models the capabilities needed for a real intelligence explosion.

  • @2945antonio
    @2945antonio29 күн бұрын

    I am a lay person on AI but watched the entire interview in two sessions - IT WAS SPARKLING! ( A especial thumbs up for the creative beer chair. Was an AI creation?).

  • @gnarfan2179
    @gnarfan21798 күн бұрын

    Dwarkesh, I very recently discovered your channel (through r/singularity, lol). You are a phenomenal interviewer, and you're great at asking probing questions. Even if I don't understand all of the jargon, I still learn a lot from listening to you and am inspired by you. I'm currently a freshman computer science major at a community college. Do you have any advice me about entering the AI research field? For instance, how do I come up with side projects that are both enlightening and impressive to potential employers? Furthermore, how do you stay driven and disciplined to continually study and practice like I'm sure you and your guests do. Sholto mentioned having studied for at least 6-8 hours every day on the weekends. I would appreciate any other advice you have for students. Thank you

  • @caseymurray7722
    @caseymurray772216 күн бұрын

    Arc of A Scythe relates an interesting future and moral conflicts that AGSI will have to overcome. The relations between different AGI will effect decision making and output at a higher AGSI level.

  • @UnDaoDu
    @UnDaoDu29 күн бұрын

    This was amazing!

  • @lesshishkin371
    @lesshishkin371Ай бұрын

    where can I read a paper about the similarity of the brain of biological organisms and the transformer architecture that he is talking about?

  • @lolololo-cx4dp

    @lolololo-cx4dp

    29 күн бұрын

    Me too, transformer math are pretty clear, "anyone" can reproduce it. I am not sure about brain's math tho.

  • @julienbaneux2099
    @julienbaneux20995 күн бұрын

    @dwarkeshpatel your “Yikes” at 58:05 had me laughing HARD.

  • @sucim
    @sucim29 күн бұрын

    Great episode! I wouldn't mind if it was even more technical though. Like explicitly talking about the embeddings, dot-products, superposition etc.

  • @berbank
    @berbank26 күн бұрын

    The association paradigm that Trenton Bricken mentions is a rich seam to mine.

  • @MaskedEngineer-kj5kt
    @MaskedEngineer-kj5kt9 күн бұрын

    Hey man, really good content. Appreciate it. Would you mind sharing the paper you guys are talking in the description?

  • @atharva__shukla
    @atharva__shuklaАй бұрын

    This was great!

  • @user-qj7ku6wi3b
    @user-qj7ku6wi3bАй бұрын

    This is incredible.

  • @friscofatseas5696
    @friscofatseas569626 күн бұрын

    Whenever I’m not doing something I find myself thinking about AI it’s my brightest hope

  • @prepthenoodles
    @prepthenoodles19 күн бұрын

    🎯 Key Takeaways for quick navigation: 03:34 *🎮 Models with long context lengths show promising potential to learn and reason through tasks, potentially outperforming human experts in certain domains like language learning and possibly in complex tasks like mastering Atari games.* 11:40 *🤔 Concerns over quadratic attention costs for long context windows may be overstated, as the costs are mitigated by other factors such as model size and the linearity of certain operations during token generation.* 15:37 *💡 While models can store and process vast amounts of information, true reasoning capabilities and human-like understanding may still be lacking, suggesting a distinction between raw information storage and higher-level reasoning processes.* 19:23 *🧠 Understanding GPT-7's model: Information flows through the model, compressed and modified at different stages, aiding in token prediction.* 20:46 *🧠 Brain analogies: Analogies between residual streams of compressed information in GPT-7 and brain processing, particularly in the cerebellum, suggest similarities in information routing.* 23:37 *🧠 Associative memory: The cerebellar circuit and attention operation closely resemble an associative memory algorithm, indicating a convergence across organisms and supporting the success of transformers like GPT-7.* 27:02 *🧠 Reasoning and pattern matching: Higher-level associations and a long working memory facilitate sophisticated queries and reasoning akin to Sherlock Holmes' deductive processes.* 39:33 *📈 Progress in AI involves simpler codebases with a focus on training rather than complex coding.* 41:46 *🧠 Interpreting and understanding failures is a significant part of AI research, requiring introspection and careful analysis.* 44:34 *💡 Effective AI research involves working backward from real-world problems and constantly iterating on ideas.* 48:08 *⏱️ Fast experimentation cycles and expanding problem-solving toolboxes characterize successful AI researchers.* 50:53 *💻 Scaling AI research requires balancing compute allocation between training runs and research programs.* 53:42 *🔄 AI augments top researchers' capabilities by speeding up experiments and providing insights, rather than replacing them entirely.* 58:45 *🔄 Synthetic data creation and evolution can lead to significant AI advancements, resembling a co-evolutionary loop with human understanding and verification functions.* 59:39 *🚀 Increased participation in AI conferences like ICML leads to faster progress in AI, akin to genetic recombination fostering advancements.* 01:02:14 *💻 Continuous orders of magnitude increase in computational power may yield diminishing returns in AI capabilities, challenging the idea of rapid AGI attainment.* 01:03:45 *📈 Despite potential diminishing returns, each incremental advancement in AI models still brings significant improvements in performance and capabilities.* 01:06:29 *💡 Despite current AI models' scale, they remain far from the complexity and efficiency of the human brain, highlighting the algorithmic overhead in AI development.* 01:14:20 *🔍 Chain-of-thought mechanism allows models to dedicate more computational resources to complex tasks, resembling adaptive compute allocation.* 01:19:23 *🧠 Investigating the interpretability of open source models is crucial for understanding their behavior.* 01:21:47 *🔄 Models communicating with each other raise questions about trust and understanding their interactions.* 01:23:20 *💡 Learning from dictionary features could provide more human-interpretable insights into AI models.* 01:24:40 *🤝 Collaboration among AI agents might precede a transition to a single, large model approach in AI firms.* 01:26:03 *📊 Future AI systems may dynamically adjust compute resources and context to specialize in different tasks.* 01:35:33 *🎯 Effective execution and experimentation are key factors contributing to progress in AI research.* 01:38:46 *🚀 Being adaptable and open-minded, especially at a young age, can lead to success in various fields.* 01:39:40 *💡 Persistence and willingness to pursue solutions to the end are crucial qualities for success in any endeavor.* 01:40:42 *🎯 Identifying high-leverage problems and pursuing them with determination can lead to impactful results.* 01:42:32 *🚧 Overcoming organizational blockers and inspiring others to push past limitations can massively increase productivity and impact.* 01:46:57 *🔍 Mentorship and collaboration with experienced individuals can accelerate learning and professional growth.* 01:48:51 *🔄 Understanding both algorithms and systems can significantly enhance effectiveness in machine learning research and engineering.* 01:57:47 *💼 Networking and mentorship play crucial roles in tech hiring, often surpassing traditional application processes.* 01:59:18 *🛠️ Job application processes aren't always straightforward; showcasing agency and world-class skills can be more impactful than following conventional paths.* 02:01:35 *🤔 The hiring process may involve biases, but designing interviews to test the right skills is essential.* 02:02:30 *🚀 Taking initiative is key in career advancement; the system won't necessarily support you, so proactiveness is crucial.* 02:04:19 *💡 Prioritizing work-life balance is valid, but seizing opportunities to excel can lead to significant career advancements.* 02:07:20 *🏅 Diligent effort and hard work can lead to becoming world-class in a field, even with intense competition.* 02:14:04 *🤔 Critiques of AI models often focus on whether features are predictive and whether higher-level associations exist beyond discrete features.* 02:16:47 *🧐 Identifying and understanding features in AI models can lead to more transparent and trustworthy outputs, aiding in tasks like code review.* 02:19:30 *💡 Understanding complex tasks in AI models involves chaining together various circuits performing basic operations, leading to unique behaviors.* 02:21:47 *🔍 Analyzing model behavior like deception requires identifying specific circuits responsible for certain actions, which can be challenging but essential for interpretability.* 02:23:32 *🛠️ Coarse-graining representations in AI models can help understand superhuman performance by decomposing complex behaviors into simpler circuits or features.* 02:24:29 *🧠 Exploring feature universality across models reveals shared features like Base64 encodings, suggesting certain fundamental aspects of learning are consistent.* 02:32:18 *🤔 Feature universality implies that certain ways of thinking and understanding the world may be shared across different intelligences, potentially reducing concerns about AI alignment.* 02:36:39 *🧠 Feature splitting in models like GPT-7 involves learning specific features for various categories, potentially improving model understanding.* 02:39:03 *💡 Understanding the weights of a model independently of activations is a challenging but essential goal for improving model interpretability.* 02:40:57 *🌳 Feature splitting allows focusing on specific features of interest, enabling more efficient exploration of semantic feature space.* 02:43:20 *🧩 Exploring subtrees of feature space in models like GPT-7 can reveal unexpected features that may not align with high-level abstractions.* 02:50:44 *🤖 Vector symbolic architectures and superposition resemble aspects of Good Old-Fashioned AI (GOFAI), offering insights into how intelligence might work in models and brains.* 02:52:39 *🕵️‍♂️ Identifying deception circuits in models like GPT-7 requires post-hoc labeling of features and exploration of feature interactions across layers.* 02:55:59 *🤔 Training GPT-7 involves finding directions that matter, similar to fitting a linear probe, but with the hope of discovering multiple directions that highlight deceptive behavior across data distributions.* 02:56:55 *🧠 Research is focused on scaling up ASL-4 models like GPT-7, with efforts divided between scaling up dictionary learning, identifying circuits, and achieving success with attention heads, aiming for progress within six months.* 02:58:23 *🔍 Publicly shared results reveal higher-level associations in GPT-7, such as those related to love and sudden changes in scenes like declarations of war, indicating deeper layers of abstraction.* 03:00:36 *🤖 Human interpretability research on AI models aims to understand and potentially ablate features to mitigate negative behaviors, considering complexities like recognizing both good and bad concepts.* 03:02:03 *🛠️ Precise tools for identifying and ablating model features offer hope for ensuring safety and reliability, contrasting with methods like reinforcement learning from human feedback which may lack precision.* 03:11:20 *🤯 Humans may seek predictability and control over their environment, preferring familiarity over surprises, reflecting on the psychology of learning and exploration.* Made with HARPA AI

  • @joshismyhandle
    @joshismyhandleАй бұрын

    Good stuff!

  • @diga4696
    @diga469627 күн бұрын

    Great to see free energy principle believers! Would love to see an interview which bridges ml/ai to fep.

  • @dnkrocks
    @dnkrocks20 күн бұрын

    WHY HASN"T GOOGLE ALGO SUGGESTED THIS TO ME EARLIER? GOOGLE IS FAILING OR WHAT? This is such a great conversation!!!!!!

  • @asavu
    @asavuАй бұрын

    Thought while watching this: What if episodic memory is just a context window with LRU semantics? Any research on that? Can we determine information usage on the forward pass?

  • @hossromani
    @hossromani28 күн бұрын

    Which camera is the pod using? Awesome setup, and which tripods - of course excellent talk

  • @diamond_s
    @diamond_s16 күн бұрын

    Sample efficiency of human brain is not just scale animals with a small fraction of brain size can start walking and fully functioning in environment within moments of birth only being trained with a few simple pattern generator patterns before birth.

  • @collins4359
    @collins4359Ай бұрын

    this is great! employees from 2 top AI companies. nice to see stuff like this.

  • @biesman5
    @biesman520 күн бұрын

    Great interview. It would be really cool to get Geroge Hotz on the podcast, he is a super interesting dude.

  • @retishe7660
    @retishe766017 күн бұрын

    Now this is an AI podcast 👏 🙌

  • @TimeLordRaps
    @TimeLordRaps28 күн бұрын

    At what point of higher dimensionality of the model projection based on the scale of the training data to parameter count would a meaningful enough denseness be associated with a lack of featureness? Would it be linked directly to the under-parameterization of the training data or the actual dimensionality of the feature?

  • @dr.mikeybee
    @dr.mikeybee27 күн бұрын

    First class content!

  • @thinkingcitizen
    @thinkingcitizenАй бұрын

    POV: Rajpoot prince talks to a surfer model and wine sommelier about AI and engineering

  • @therainman7777

    @therainman7777

    Ай бұрын

    Dumb

  • @thinkingcitizen

    @thinkingcitizen

    Ай бұрын

    @@therainman7777 this comment was generated by a video-to-text AI model

  • @schwajj

    @schwajj

    27 күн бұрын

    @@thinkingcitizenwhy not say so in the first place? Which model? What prompt?

  • @skoto8219

    @skoto8219

    27 күн бұрын

    I really was thinking to myself “these three guys have no business being this good looking while having this kind of discussion”

  • @dr.mikeybee
    @dr.mikeybee17 күн бұрын

    I hadn't thought of the transduction in transformers as learning from content, but it's quite so. The question is what is being learned? What is the new abstract representation? It seems to be gaining geometric information about the model. It produces a context signature that touches points on the manifold. But what exactly is that doing? I can intuit hierarchical abstract structure and decision selection being created in FFNNs, but what is happening in the attention heads?

  • @alexmolyneux816
    @alexmolyneux81622 күн бұрын

    QUESTION FOR TRENTON: I just had a question on ‘intelligence is just associations’ - i dont think this was explored fully. Its an interesting idea. Its not clear to me how you move from associations to problem solving and deduction. I can see how you learn a bunch of associations from corpus A, which allows much easier learning of corpus B. But how does this translate to, ‘heres everything we know about physics, please create a unified theory of everything’. I could perhaps see that you're now projecting all these rich learnt associations at a new thought vector. It feels that there is a difference between learning these associations and problem solving and finding solutions though. A solution to me seems like a way to minimise entropy with minimal energy. With a particular effort you unlock a lot of order. First would be interesting if you agree or have thoughts on that, because really I think we care a lot about the solving aspect ‘please solve fusion’ for example. And secondly, would be interesting to know if you can see a mapping between associations and problem solving, or you see them as one and the same.

  • @skoto8219
    @skoto821928 күн бұрын

    18:07 with tangerine trees and marmalade skies

  • @priyeshsrivastava8025
    @priyeshsrivastava802529 күн бұрын

    We should just caption all papers mentioned in this podcast.

  • @colindocherty3254
    @colindocherty325429 күн бұрын

    some of the world's most powerful ideas in just 3 hours

  • @NikhilKumar-fo5on
    @NikhilKumar-fo5on28 күн бұрын

    i like that *Interpretability* shirt!

  • @TimeLordRaps
    @TimeLordRaps28 күн бұрын

    So the argument is to only pursue paths towards possible negative features, but the curse of dimensionality so aptly named may indicate that there are features as Dwarkesh hinted at that are require an amount of dimensionality that may be incomprehensible to a probe first approach, which hints to me that the only solution is a weights first approach because just trying to probe for things you think you'll find may help, but without the "black hole" space in the feature set it there would be no way of detecting the need for probing. Also this doesn't account for the impact of intentional feature manipulation once these models realize what's happening.

  • @DRKSTRN
    @DRKSTRNАй бұрын

    Saying it's all just associations ignores the sheer complexity of that statement. Comparisons are the basis of logic and reasoning. And associations are simply a pathed comparison that can be made/known. For example: 'A' relays to all letters starting with a and use such as a vowel if playing scrabble/wordle. This also ignores the surrounding associations/comparisons/pathing that allow for this framing to be possible in the first place. Thus, finding some seemingly common sense concept that connects to a large number of parts is the discovery of something that is beyond general. But likewise is the curse of logic. It doesn't matter how difficult it was to formalize that 10 step hierachal planning process over a course of a month. If it is logical, it just makes sense to outside oberservers. A to B.

  • @ashh3051
    @ashh305124 күн бұрын

    Almost didn’t watch because of the clickbait title. But it was worth it.

  • @segelmark
    @segelmark28 күн бұрын

    Another great episode ☺️ Would you mind starting a separate channel called DwarkeshClips or something and putting the short clips there, so it is easier to see when you have a full episode like this one 🙃

  • @HarpreetSingh-xg2zm

    @HarpreetSingh-xg2zm

    27 күн бұрын

    Probably need more subscribers before he can branch out this channel. Right now he is trying to funnel as many people to main channel

  • @davtak
    @davtak22 күн бұрын

    Thanks!

  • @darylallen2485
    @darylallen2485Ай бұрын

    He's a Tech Bro!

Келесі