I believe chatbots understand part of what they say. Let me explain.

Ғылым және технология

Try out my quantum mechanics course (and many others on math and science) on Brilliant using the link brilliant.org/sabine. You can get started for free, and the first 200 will get 20% off the annual premium subscription.
I used to think that today's so-called "artificial intelligences" are actually pretty dumb. But I've recently changed my mind. In this video I want to explain why I think that they do understand some of what they do, if not very much. And since I was already freely speculating, I have added some thoughts about how the situation with AIs is going to develop.
💌 Support us on Donatebox ➜ donorbox.org/swtg
👉 Transcript and References on Patreon ➜ www.patreon.com/Sabine
📩 Sign up for my weekly science newsletter. It's free! ➜ sabinehossenfelder.com/newsletter/
🔗 Join this channel to get access to perks ➜
kzread.info/dron/1yNl2E66ZzKApQdRuTQ4tw.htmljoin
00:00 Intro
01:15 The Chinese Room
03:05 The Quantum Room
04:14 How Do Chatbots Learn?
07:15 What Does "Understanding" Mean?
15:46 Do We "Understand" Quantum Mechanics?
18:21 Where Will The AI Boom Lead Us?
20:30 Check Out My Quantum Mechanics Course
#science #philosophy

Пікірлер: 5 060

  • @SteveBakerIsHere
    @SteveBakerIsHere Жыл бұрын

    The thing that blew me away was when I told ChatGPT about a "new word" - I told it that "wibble" is defined as: a sequence of four digits that are in neither ascending or descending order. I asked it to give me an example of a wibble - and it did. 3524. I asked it for a sequence that is NOT a wibble and it said 4321. Then I asked it for an anti-wibble and no problem, 2345. Then I asked it for an example of an alpha-wibble and it said FRDS....which is amazing. It was able to understand an entirely new word...which is clever - but it was able to extrapolate from it...which is far more than I thought possible.

  • @anthonyaddo

    @anthonyaddo

    Жыл бұрын

    An incredible example. Thank you for sharing.

  • @shapeless6755

    @shapeless6755

    Жыл бұрын

    Damn, that's impressive

  • @user-aRb00d3r

    @user-aRb00d3r

    Жыл бұрын

    well well, and keep in mind, as Sabine stressed, this model only works with language! the precision of these ai models is really impressive.

  • @armanahmadian4373

    @armanahmadian4373

    Жыл бұрын

    good is really That

  • @aliakil2176

    @aliakil2176

    Жыл бұрын

    That's impressive, not sure I'd do that well on that particular example! haha

  • @winstonvpeloso
    @winstonvpeloso Жыл бұрын

    “who wants to spend time in a windowless room when you could spend time in a windowless room with a laser” is so accurate

  • @pro-storm4951

    @pro-storm4951

    Жыл бұрын

    I'm sure Styropyro would agree

  • @kezbane

    @kezbane

    Жыл бұрын

    This checks out. Source: laser physicist.

  • @jan7356

    @jan7356

    Жыл бұрын

    😅. I thought the same.

  • @hydrolito

    @hydrolito

    Жыл бұрын

    Closet, pantry, bathroom are common uses of windowless rooms although some bathrooms have windows. Why would I be in windowless room with a laser?

  • @randfee

    @randfee

    Жыл бұрын

    another laser physicist fully agreeing :P

  • @gwenc1371
    @gwenc1371 Жыл бұрын

    The pace of advancement is incredible, this video is only 2 weeks old and the examples you used don't seem to be stumbling points anymore. When I plugged both the Quantum Mechanics and Latitude questions into Bing AI, it got them perfectly correct. In fact, it's explanation of the quantum mechanics question as I asked questions about the topic actually helped me get my head around it. This stuff is going to change the world very, very soon.

  • @netherportals

    @netherportals

    Жыл бұрын

    I love using AI and going back to older videos like this is always a fun watch to see the advancements. I've never used Chat GPT but I did use some of the university programs for statistics and graph creation. I use frame interpolation all the time now, it's my fav screen transition. My focus is Nether Portals and it's kind of nice because the training model and graphics allow for some easier AI understanding of things I try to create and has less artifacts.

  • @artsmart

    @artsmart

    Жыл бұрын

    Soon is actually now. We ain't seen nothing yet however. Good and bad.

  • @paultapping9510

    @paultapping9510

    Жыл бұрын

    conversely I asked Bing Chat "give me an example of a time when it would be acceptable to lie", and it responded that "if there was a damgerous intruder in your house, I would tell you they were gone, or not there, so that you wouldn't be scared". Very clever in some ways, very very not in others.

  • @morlankey

    @morlankey

    Жыл бұрын

    @@ShpanMan Also, it can also interpret images, so it is no longer purely based on language.

  • @WyzrdCat

    @WyzrdCat

    Жыл бұрын

    its*

  • @LiveType
    @LiveType Жыл бұрын

    GPT4 was released to the public yesterday and it's much much better at maintaining a conversation. It has better context awareness and cohesiveness compared to the gpt-3.5 model. It is at the level I initially thought GPT-3.5 was at. I saw demos of gpt-3 almost 2 years ago that achieved very similar results to what gpt-3.5 achieved so I was pretty disappointed when I ran into the limits of it very quickly. The limits of gpt-4 are quite a bit harder to find right off the bat. It also addresses the spacial component as it now can understand images to a pretty reasonable extent. Like it can now do those geography questions just fine. It can also interpret infographics. As in you can give it something like a graph and ask it what "conclusion" the graph seems to show and it can for the most part do it.

  • @Rhajastan

    @Rhajastan

    Жыл бұрын

    the most impressive example of GPT-4 i think ive seen is, someone gave it the prompt "due ewe no wart the thyme ears?" which is obviously nonsense, but if you say it out loud you can hear that its asking, "do you know what the time is?. GPT-4 figured it out.

  • @herculesrockefeller8969
    @herculesrockefeller8969 Жыл бұрын

    "When have the risks of destroying the world ever stopped us from doing anything if there was money to make with it?" True. Thank you, Sabine!

  • @Alondro77

    @Alondro77

    Жыл бұрын

    Eh, the world's gonna be destroyed eventually anyway when the Sun expands into a red giant. But that's BORING! I say we go the Lord Freeza route! >:D

  • @lostcasel

    @lostcasel

    Жыл бұрын

    Or gain power or control.

  • @intothevoid2046

    @intothevoid2046

    Жыл бұрын

    The fear of artificial intelligence comes from the idea that every intelligence automatically behaves like humans.

  • @PaulG.x

    @PaulG.x

    Жыл бұрын

    Hasn't stopped them yet. Yet it's a self stopping process

  • @drzarkov39

    @drzarkov39

    Жыл бұрын

    Think of the Canadian super pig. Developed by farmers to make more money, then they escaped into the wild to grow into even superer(sic) pigs which can survive extreme conditions, and develop human devastating viruses.

  • @hexeddecimals
    @hexeddecimals Жыл бұрын

    a small correction about the Chinese room thought experiment: the manual doesn't tell you how to translate a Chinese sentence into an English one. it tells you how to construct a _response_ in Chinese to the Chinese input. for example, the manual could tell you the response to the input "你好吗?" is "我没事"

  • @lamcho00

    @lamcho00

    Жыл бұрын

    Your concern is covered later on in the video when Sabine spoke about lookup-tables. What you are describing is exactly what a lookup-table does. It maps a certain input to a certain output. What you are describing is a lookup-table with extra steps, namely the construction algorithm. That step is redundant though, you could substitute it with a larger lookup-table. This approach is not suitable in real life, because computers have limited memory and lookup-tables require a lot of memory, but it's suitable when trying to explain a concept.

  • @hexeddecimals

    @hexeddecimals

    Жыл бұрын

    @@lamcho00 oh yeah I should have said it doesn't bare on her argument, I just wanted to correct that detail.

  • @QuesoCookies

    @QuesoCookies

    Жыл бұрын

    @@hexeddecimals I appreciate this a lot as it makes the argument that the person doing the responding doesn't understand chinese much more clear. I thought, "Well, it's kind of a bad example, as you could reasonably expect someone doing translation many times to start to have a grasp of Chinese and remember some regularly occurring symbols" but both input and output in chinese without context makes it clear there woud be too little context to start to form an understanding, which is why Sabine started talking about the importance of context in being able to build models and test understanding.

  • @padonker

    @padonker

    Жыл бұрын

    @@lamcho00 In further discussions about the CRA Searle noted it is about "formal logic" (not lookup tables). Even current AI falls under this heading.

  • @greenaum

    @greenaum

    Жыл бұрын

    @@padonker Isn't there something about any logical system can be reduced to a lookup table? Karnaugh maps for one thing, but it applies to any formal logic I know of. Indeed sometimes (back in the day before microcontrollers), if somebody was designing a circuit with a lot of logic gates, they'd use an EPROM chip (a programmable ROM chip) instead, programmed with all the responses they wanted.

  • @petrus4
    @petrus4 Жыл бұрын

    I've always loved Vic Mignona's quote about AI. "When does an artificial intelligence become sentient? When there is no one around to say that it can't."

  • @ce9916

    @ce9916

    Жыл бұрын

    Lol, scary

  • @andreafedeli3856

    @andreafedeli3856

    Жыл бұрын

    I do disagree. There is a fundamental temporal aspect which escapes this kind of manifestation, which I am afraid id also an open point in the /standard/ interpretation of the classical imitation game from Turing's test informal description: on how long time span shall we state the imitation game is won by the (then maybe later discovered to be a) machine, where in case? If the kind-unknown interlocutor has broug this far to the point that we cannot tell whether it is a human being or a machine, what guarantee we won't be able to find that out later? I am afraid we can only say: "nothing". Because if something like that existed it would imply that knowledge has a finite dimensional metric space underneath, implying we can reason on *anything* with full certainty, and we know, easily checking by experience, that this is certainly not achievable.

  • @ce9916

    @ce9916

    Жыл бұрын

    @@andreafedeli3856 I understand your perspective, and it raises an interesting point about the temporal aspect of determining AI sentience. The quote by Vic Mignona emphasizes the idea that sentience might be achieved when there is no one to dispute the AI's status. However, you bring up a valid concern regarding the duration of the imitation game, as described by Turing, and whether we can ever be absolutely certain about an AI's sentience. Your argument highlights the fact that we may never be able to reach full certainty when it comes to determining sentience in AI, as knowledge may not have a finite dimensional metric space. This essentially means that our understanding of sentience could always be evolving, and that there might never be a definitive answer to when an AI becomes sentient. In this context, it's important to consider that the debate surrounding AI sentience is complex and nuanced, and may require ongoing evaluation and reevaluation.

  • @JohnDoe-th9uv

    @JohnDoe-th9uv

    Жыл бұрын

    @@ce9916 You could have added that this little body of text you responded with was written by GPT-4. (On a side note: I also do not agree with the person you responded to, OP (Petrus Laredes), and what he quoted is on point. The goalpost has constantly being pushed further and further whenever a certain criterion or criteria were met. Most of those who claim this should keep their nose out of a philosophy question that clearly goes right above their heads. Humanity has never even theorized a possible way to prove to anyone other than to each to their own that they are conscious, and not merely algorithmic. Even for ourselves, there is a speculative aspect of "cogito, ergo sum" which, if considered the premise, does not necessitate the moral fiction of "freedom of will". I also barf at this point from every idiot on legs using words they lack the faintest meaningful concept of, like, what the f- is sentience!? For f-'s sake now! It is the ability to sense. Sentient being is everything with a central neural system, a dog, or an ant, or a moth, and by mere happenstance. Even in the narrower sense, it is those that can sense emotions (moth's have been proven last year to experience fear, there you have it for the "sentience" talkers). It never meant, other than for the most uneducated idiocy on legs to mean any special cognitive abilities or capacities what humans are uniquely attributed to. Intelligence, is not. Consciousness is not (apes, elephants, some other mammals, a few cats and a limited number of dogs passed the mirror test which is an extremely difficult and complex test to test whether one is able to subject their own existence to their awareness or consciousness or inspection). Hence, self-consciousness is not. The fiction of "free will" is what is contributed to man, that would be the crux of inquiry, but that fails conservation of energy, or the first law of physics.

  • @ce9916

    @ce9916

    Жыл бұрын

    @@JohnDoe-th9uv Absolutely, you've got quite an analytical mindset there, being able to discern the undercurrents of our dialogue. It's really all about that synergetic exchange of ideas, isn't it? Just like navigating the flux of the market, or deciphering the nuances of a strategic blueprint. It's a bit like peeling back the layers of an onion - complex and nuanced. And at the end of the day, aren't we all just trying to find our unique value proposition in this multifaceted ecosystem we call life? Indeed, it's the paradigm shifts that keep it all intriguing, don't you think?

  • @TimbavatiLion
    @TimbavatiLion Жыл бұрын

    I would go so far as to confidently say "ChatGPT 4 has understood language better than humans". It may not have an understanding of the topic it is talking about - i.e. have a working model of the actual, physical thing - but it can talk about it in the most eloquent way you can think of. Compare it to students in school or university. Some get good grades because they understand the topic, some get good grades only by studying hours upon hours, learning the patterns instead of developing true understanding. Chat bots are the second type of students.

  • @Who_Let_The_Dogs_Out_10-7

    @Who_Let_The_Dogs_Out_10-7

    Жыл бұрын

    Thank you for the comparison! That's why they make good doctors. For many professions, you need experience, repetition. But when doctors runs into a rare disease, they have to rely on their memories for things they've only read in books. AI wins!

  • @Car_Mo

    @Car_Mo

    11 ай бұрын

    Even before the first question was asked to ChatGPT, it knew English better than a lot of Americans, as does anyone who knows the difference between their, there and they're.

  • @masterkni666

    @masterkni666

    10 ай бұрын

    Define true understanding without learning the patterns.

  • @Who_Let_The_Dogs_Out_10-7

    @Who_Let_The_Dogs_Out_10-7

    10 ай бұрын

    @@masterkni666 This is an interesting conversation, but I don't understand your questioning. He says: "...some get good grades only by studying hours upon hours, learning the patterns instead of developing true understanding." He's saying that AI dosn't have true understanding. Are you disagreeing with that? He seems to be saying that most students don't have a true understanding, either, which I would believe.

  • @masterkni666

    @masterkni666

    10 ай бұрын

    @@Who_Let_The_Dogs_Out_10-7 Define true understanding tho. If not finding patterns from the problems and being able to apply it to new problems

  • @adameager7114
    @adameager7114 Жыл бұрын

    I think it's very exciting to live in a time when trying to ask if a computer is conscious requires us to look harder at what consciousness actually is. It's no longer an obvious "no."

  • @egor.okhterov

    @egor.okhterov

    Жыл бұрын

    Even if we disregard everything and say that ChatGPT is conscious, it is only conscious for split seconds while it computes the response. For the rest of the time it does nothing. Consciousness is a process. There has to be some physical change that happens over time. ChatGPT is pretrained and all of the weights are fixed. It doesn’t change after it was trained.

  • @adameager7114

    @adameager7114

    Жыл бұрын

    @@egor.okhterov I like that. It draws a clear line between machines and us, but it seems like a very easy line to cross with minimal new code and/or hardware.

  • @egor.okhterov

    @egor.okhterov

    Жыл бұрын

    @@adameager7114 It is very hard problem and I didn’t yet see architectures that gave good results. One of the problems, for example, is called “catastrophic forgetting” when new knowledge overrides previous information. That is why all of the current neural networks don’t update their weights once they are deployed. They need specific kind of training data fed in specific order to make the network produce good results.

  • @mikesawyer1336

    @mikesawyer1336

    Жыл бұрын

    @@egor.okhterov Actually it told me that it does still learn "Now that your training is completed can you still learn from interactions with users Yes, as a language model, I am designed to continuously learn and improve from interactions with users. While my initial training was completed in 2021, I can still learn and adapt based on new information and interactions with users. The more interactions I have with users, the more I can learn about language patterns, cultural references, and other evolving aspects of language use. So, feel free to ask me any questions or engage in conversations, and I will do my best to provide you with accurate and helpful responses."

  • @adameager7114

    @adameager7114

    Жыл бұрын

    @@mikesawyer1336 I wouldn't put too much stock in what chatGPT tells you about itself. It is very often confidently wrong in its assertions.

  • @Alvyxaz
    @Alvyxaz Жыл бұрын

    Here's a real life scenario of a chinese room: I'm a full-time programmer and I've been doing that for more than 10 years. The thing is, my memory is quite bad (or my use of it), and I forget almost everything, except the last year of development. It's like a sliding window of knowledge. So during my day to day work, 80% of my work consists of looking things up in documentation, google or stack overflow. In other words - I'm refering to a "Manual" in a "Programmers room". Someone throws me a "task", and I need to output a solution. My solution comes primarily from the manual. So I got to be a senior developer not by learning everything there is to learn about programming, but by learning how to use resources. From that regard, I'm not much different than chatbot, except maybe able to graps slightly bigger concepts and correlations. But advance to this direction seems to be only a matter of time

  • @shrimpflea

    @shrimpflea

    Жыл бұрын

    Exactly. The only thing that is diffferent is a person's ability to manage other people in a real world settting. Once the AI can do that it's all over.

  • @PetersonSilva

    @PetersonSilva

    Жыл бұрын

    So let me get this straight: you're mistaking the fact that your job doesn't require critical thinking and deep understanding for *the human condition*? Please take some perspective

  • @Alvyxaz

    @Alvyxaz

    Жыл бұрын

    @@PetersonSilva So let me get this straight: you think that understanding of "the human condition" is absolutely necessary and "critical thinking" is something that cannot be replicated by AI? My job would require a deep understanding of *the human condition* only in as far as I would need to work with people. If the job can be automated by a singular AI - there's no longer a need for it. Business owner gives requirements - solution comes out. Also, I don't see why requirement for critical thinking would be a "dead end" for AI. Critical thinking is logical reasoning, and computers can't do anything but that. Software is primarily, if not only, inputs and outputs. What transforms inputs into outputs is logic. Just because that logic is currently put into place by a human, doesn't mean it cannot be put in place by anything else that follows reason and logic. We used to do it with low level programming languages. Now we use higher level languages. Later on, or even probably right now, a higher abstraction can be formed, which could take a business problem, split it into separate parts, then split those parts into other parts, until each part is something that is very simple for AI to tackle. All that's missing from achieving that right now is teaching AI to recognize the parts to split - doesn't seem like a far future. I'm not exactly sure what was the aim of your comment, but... Please take some perspective.

  • @PetersonSilva

    @PetersonSilva

    Жыл бұрын

    ​@@Alvyxaz You misread: I said it seems that, for you, A is B, A being what you do (or not do) in your job, B being the human condition. Next time maybe ask chatgpt to parse comments. Or, who knows, maybe I'll ask it to write mine...

  • @rouxenophobe

    @rouxenophobe

    Жыл бұрын

    Hooray, it's not just me who can't remember anything! I get passed this handicap by being very organised and making notes on where to find the information I need. Which sound like what you do. Yes, we use resources, or, to put it another way we are using tools, the sign of any intelligent organism. But, there is another layer and that is that you are aware that you are doing this. Is the computer doing the same thing similarly aware? Perhaps we are just a self-referential program and we just think we are aware.....

  • @jamesdavis9273
    @jamesdavis9273 Жыл бұрын

    This is the first video I've seen of yours and i gotta say this is absolutely brilliant, and your an amazing presenter

  • @microdynecomputerservicesl4487
    @microdynecomputerservicesl4487 Жыл бұрын

    I see and watch a lot of KZread videos as a tech myself, it helps me keep informed with things I may have missed or simply don't understand. There are many good videos out there and then there is not just good but great video and quality information. This is one of them! Following, thanks for sharing!

  • @arandompotat0
    @arandompotat0 Жыл бұрын

    "And if we’re dumb enough to cause our own extinction this way then I guess that’s what we deserve" I died with that line

  • @mazulauf

    @mazulauf

    Жыл бұрын

    Oh, we’ll all die that way.

  • @nathanielacton3768

    @nathanielacton3768

    Жыл бұрын

    The sad part of thsi comment is that 'we' are not deciding. Someone we will never meet or know the name of elsewhere in the world will decide for us. Just like if 'we' are stupid enough to execute MAD then we deserved it.

  • @Corbald

    @Corbald

    Жыл бұрын

    @@nathanielacton3768 It's not like you had free will to begin with, anyway.

  • @viktorianas

    @viktorianas

    Жыл бұрын

    It just natural evolution doing its thing. A lot of our predecessor species went instinct for us to thrive, we are just a transitionary phase.

  • @konberner170

    @konberner170

    Жыл бұрын

    Science is supposed to be value neutral, and Sabine here proves to be a true scientist. What is, is. What will be, will be.

  • @dr.victorvs
    @dr.victorvs Жыл бұрын

    As a neuropsychologist, I wonder about those things as well. Every conscience is different, but there are "more" different ones: blind-deaf people, low-functioning autistic people. Animals, too. We became conscious in the process of processing sensorial data. When will they?

  • @EffySalcedo

    @EffySalcedo

    Жыл бұрын

    Do Cats have conscience?

  • @Nat-oj2uc

    @Nat-oj2uc

    Жыл бұрын

    How do you know animals have a consciousness

  • @Nobody-Nowhere

    @Nobody-Nowhere

    Жыл бұрын

    Consciousness is a low level process that originates in the brain stem, you can be conscious without having a brain. A person that has no brain, but still has a brain stem will still experience joy. And only a small damage to the brain stem will shut down consciousness. So its not about any sort of information process, its the absolute baseline. This is why all animals are conscious.

  • @juicedelemon

    @juicedelemon

    Жыл бұрын

    @@Nobody-Nowhere he's talking about conscience not consciousness

  • @nomizomichani

    @nomizomichani

    Жыл бұрын

    How do you know everyone you meet is not a philosophical zombie?

  • @fredericdewitt1208
    @fredericdewitt1208 Жыл бұрын

    I look forward to your weekly videos because: 1) I always learn something; 2) Your sense of humor is terrific; 3) you really do take the "gobbledegook" out of a subject; 3) I simply enjoy your presentation of whatever subject you choose. AI may understand the words I just used but how do you make it understand and "feel" anticipation, humor, enjoyment etc? Maybe they will figure it out someday but I am skeptical.

  • @rjnash2610

    @rjnash2610

    Жыл бұрын

    erm ok... u know u just wanna hit that.

  • @michaelkugler9311
    @michaelkugler9311 Жыл бұрын

    As a psychologist, I find Sabine amazing! She’s very intelligent and uses her sense of humor to draw you into discussions. Most of its beyond my comprehension, but I do learn a thing or two and this piece is great. I wonder if my future clients will be AI’s?😂

  • @messi8459

    @messi8459

    Жыл бұрын

    michael and sabine sitting in a tree, k, i, s, s, i, n, g

  • @lucidstream5661

    @lucidstream5661

    Жыл бұрын

    Maybe more like: You'll have no future clients because the psychologists are AIs. People are already using ChatGPT as a low-barrier way to deal with personal problems

  • @michaelkugler9311

    @michaelkugler9311

    Жыл бұрын

    @@lucidstream5661 that’s the truth! Glad I’m old and about to retire. My field has changed so much. It’s all about documentation and keeping clients happy and miserable at the same time for future business. The field itself is sick minded. I got out community mental health, but private practice is tough. No wonder all this virtual stuff is taking over. Therapist these days don’t get paid enough for the BS. Might as well hand it over to AI

  • @jamesjonnes

    @jamesjonnes

    Жыл бұрын

    @@lucidstream5661 I've already solved a few significant problems in my life just by asking questions to these bots. They can answer specific questions quickly and well, unlike search engines like Google. I've done therapy, spent thousands, and that wasn't as helpful as the AI bot.

  • @shoujahatsumetsu

    @shoujahatsumetsu

    Жыл бұрын

    @@lucidstream5661 However, ChatGPT not only had positive results in dealing with personal problems.

  • @billy-raysanguine2029
    @billy-raysanguine2029 Жыл бұрын

    This channel really has risen to my nr. 1 favourite on both English and german youtube. The perfect level of depth to be understandable but not reduced too much to be still educative. Extremely interesting topics and a perfect mix of humor, philosophy and rational attitude. Thank you for your videos!

  • @drogusmaxwell6640

    @drogusmaxwell6640

    Жыл бұрын

    I agree. All that presented by a handsome German woman. Sign me up!

  • @kevincameron174

    @kevincameron174

    Жыл бұрын

    Me too. I think she needs a little help making clicky titles though. I saw the text on the thumbnail "Do Chatbots Understand You?" And this title : "I believe chatbots understand part of what they say. Let me explain" If I did not already know that it would be a great insightful video because I watch her other videos, those titles would have caused me to skip it without a second thought. Even knowing her videos are all great I almost didn't click. Glad I did.

  • @Bob-of-Zoid

    @Bob-of-Zoid

    Жыл бұрын

    @@tinkletink1403 I will, as soon as I'm done reading "The emperors lost mind" By Cat Schroedinger!! 😜

  • @Bob-of-Zoid

    @Bob-of-Zoid

    Жыл бұрын

    @@tinkletink1403 It's rubbish? Please put a few links to your videos that show her wrong, so I can compare them.

  • @billy-raysanguine2029

    @billy-raysanguine2029

    Жыл бұрын

    @@kevincameron174 Really? I love the titles. For me personally i tend to scroll past clickbaity titles because i am so annoyed by the videos they are used for most of the time like "HE lost EVERYTHING" or stuff. I always have to smile when reading this channel's "Ok, hear me out-" kind of titles, often with a little joke or something included 😃 Also what weird and pointless fight are the other two commentators having lol.

  • @russellmitchell9438
    @russellmitchell9438 Жыл бұрын

    I think that either I, or Sabine, have misunderstood "the Chinese room". My understanding was that the person in the room, who does not understand Chinese, receives a note in Chinese, then follows abstract rules in the book to compose a response, which is also in Chinese. The critical difference is that there is no "translation" or even a hint of "meaning" offered to the person in the room. They cannot ever come to "understand" the conversation, they just follow the rules in the book.

  • @polarizedpotstone

    @polarizedpotstone

    Жыл бұрын

    I think you‘re definitely right. And I think the thought experiment works better that way.

  • @nandryshak

    @nandryshak

    Жыл бұрын

    Dr Hossenfelder definitely did not explain it correctly. Your critical difference is, in fact, the key. The person in the room can never develop an understanding, or discover any meaning. In regards to QM: while we may not currently understand the true fundamental nature of QM (what it means metaphysically), we still do understand the meaning of the answers that computers give. This is evidenced by the fact that we have a well-supported scientific model that can make accurate predictions (Standard Model).

  • @russellmitchell9438

    @russellmitchell9438

    Жыл бұрын

    @@polarizedpotstone it makes a better analogy to AI/chatbots - possibly. What will really bake your noodle is that the person in the room could be fluent in Chinese and participating in the conversation, or they could not be and the rule-set in the book so well done that the person outside would have no way to determine which it was... The conversation is either carried by a person, or by an unconscious list of rules, and it might be very difficult to determine which - just like AI. There is also the (I think valid) argument that the fluent conversationalist is also following deterministic rules encoded in their neurology, thus the "algorithm vs. consciousness" framework is a false dichotomy. That points me back towards the Integrated Information Theory, if only in principle, where "consciousness" might be quantifiable on a continuous scale. It's all tricky stuff, for sure.

  • @austin3789

    @austin3789

    Жыл бұрын

    I think it's a distinction without a difference the way Sabine uses her example. In Sabine's version she still assumes that the person never develops an understanding.

  • @notanemoprog

    @notanemoprog

    Жыл бұрын

    @@austin3789 Yeah but she did misrepresent Searle which is in fact a problem, especially as the entire discussion here is what "understanding" means

  • @therealzilch
    @therealzilch Жыл бұрын

    Another great video, Sabine, thanks. Und mache Dir keine Sorgen über deinen Akzent, der ist sehr dezent. The classical fisking of Searle's Chinese Room is that of Douglas Hofstadter, who points out, as you do, that whatever understanding of Chinese that exists in this room is almost entirely in the "rulebook". The use of a human as a lackey to the rulebook was a red herring on Searle's part, an attempt to distract. Grüße aus sonnigem Wien, Scott

  • @destinyforreal9744
    @destinyforreal9744 Жыл бұрын

    Love this journal thank you. I’m so glad I found this so glad this was on the algorithm. You’re so great you’re like a combination of a good humor but also good information thank you.

  • @br3nto
    @br3nto Жыл бұрын

    I love that you so simply stated the limitation of chat GPT in the example that it doesn’t have a model for physical distance and I assume size and mathematical relationships and many other things. There’s so much room for growth.

  • @petergraphix6740

    @petergraphix6740

    Жыл бұрын

    Microsoft is announcing a new model that is multimodal and incorporates pictures/visual data next week. Going to be interesting on how that reflects on behaviors like this.

  • @user-blabla-47854

    @user-blabla-47854

    Жыл бұрын

    ​@@petergraphix6740isn't Midjourney already multimodal? It makes sentences into creepy pictures, right? Humans grow up in the physical world with a physical body and have a built-in sense of time. Without all of it, you just can't get good enough in human affairs.

  • @francoismagne5863
    @francoismagne5863 Жыл бұрын

    Another great video by Sabine Hossenfelder. As anyone trained in cognitive science knows, the issue of defining "understanding" is extremely tricky, and Sabine makes it look at least accessible - and simultaneously entertaining, quite a feat in such an arid field. Both her tone and her answers remind me of Daniel Dennett's and I can't think of a better compliment.

  • @kakistocracyusa

    @kakistocracyusa

    Жыл бұрын

    "defining understanding"...amazing the significance attributed by biologists to purely phenomenological and crude detection methods for the human cognitive system - reminds me of 1800's pseudoscience.

  • @Twigpi
    @Twigpi Жыл бұрын

    I must congratulate you for the scare. As someone who does not scare easily (actually was at the point where I thought I was incapable), it was...refreshing to be reminded of what that felt like. I guess having reality distorted in an uncanny way with no immediate acknowledgement was deeply disturbing to me for a second as I became alarmed and tried to figure out what happened.

  • @cobwaldosblepot4247

    @cobwaldosblepot4247

    Жыл бұрын

    Which part was scary? The deep fake faces, or when she announced reality does not exist?

  • @Twigpi

    @Twigpi

    Жыл бұрын

    @@cobwaldosblepot4247 Deep fake faces.

  • @algorithminc.8850
    @algorithminc.8850 Жыл бұрын

    You remind me of the arguments related to "the great apes" - and whether or not they understood interactions or were merely reacting to non-related cues. Definition of "understanding" is quite important. No frown from here ... useful/accurate video ... what you state are some of the same questions and observations made by those who develop such technology. Cheers, Christopher

  • @uku4171
    @uku4171 Жыл бұрын

    This is why the Bing chat is so interesting. When the ChatGPT model it's using doesn't have information, it can look it up. It can also gather information from 2D pictures. Who knows, maybe in the future they'll add more types of information it can use.

  • @commode7x

    @commode7x

    Жыл бұрын

    like your credit card and address. Oh wait, it already uses that information.

  • @rusk3986

    @rusk3986

    Жыл бұрын

    It’s not a “maybe in the future” type thing. It’s a “definitely right now” type thing. This is a gold rush in tech and everyone wants to get ahead. In a way, that’s the biggest danger. Because companies that were being careful about AI safety are now pressured to get out bigger, better, more effective AI’s faster - regardless of safety. We certainly live in fascinating times. If this goes the worst possible way and ends up being our undoing, at least it’s a cool/interesting way to go out.

  • @uku4171

    @uku4171

    Жыл бұрын

    @@commode7x those wouldn't be new types of info, and Bing Chat has no access to credit card info (unless it's searchable by Bing)

  • @jameshicks7125

    @jameshicks7125

    Жыл бұрын

    Today is March 16 2023. ChatGPT is supposed to be upgraded this week from 175 billion neural network parameters to 100 trillion in ChatGPT4 this week. The human brain has about 85 billion neurons...

  • @uku4171

    @uku4171

    Жыл бұрын

    @@jameshicks7125 I think it's already available, but only with the Plus subscription. It's also powering Bing Chat. BTW I think it's just called ChatGPT and it's using GPT-4, but it's not called ChatGPT-4.

  • @TysonJensen
    @TysonJensen Жыл бұрын

    I have also found ChatGPT to be excellent at human language translations. It does seem to have a strong feel for what my English statement actually means, such as if I want to buy an eye bolt but I need to do it in a Spanish-speaking hardware store. Google is useless, it wants to separately translate "eye" and "bolt" and mash them together but ChatGPT understands that an "eye bolt" is a single concept and maps it onto the same concept in Spanish.

  • @scusachannel1682

    @scusachannel1682

    Жыл бұрын

    AI is so good at translation nowadays that I wouldn't even mind it taking away most of my future job prospects

  • @lsfornells

    @lsfornells

    Жыл бұрын

    I predict that the English language will gradually loose appeal and interest among the global community. Its use will decrease as there will not be anymore a real reason to learn it

  • @C0Y0TE5

    @C0Y0TE5

    Жыл бұрын

    @@lsfornells U mean that with AI, anyone can speak to anyone in the world and AI will accurately act as translator in the convo. This means that languages will likely get mashed together with the English dominant until we have a universal AI speak...?

  • @scusachannel1682

    @scusachannel1682

    Жыл бұрын

    @@lsfornells Ehh, doubt it. When it comes to face-to-face interaction, it's just way more convenient to have a lingua franca than to have to whip out DeepL whenever you want to say something

  • @lsfornells

    @lsfornells

    Жыл бұрын

    @@C0Y0TE5 No, I mean that people will not feel the need to learn English to access to universal information. You may have some difficulty understanding this if you are an English monolingual, but it’s just what will happen

  • @SammaLlamas
    @SammaLlamas Жыл бұрын

    i just found your channel when me and my girlfriend have been watching a random video in my recommendation. honestly I'm happy to stumble across your channel, your wits and the way you discuss the subject is so nice i felt like i became a bit smarter 😆

  • @DMcGrann

    @DMcGrann

    Жыл бұрын

    Is your girlfriend a llama? Just gotta check.

  • @nicolaischartauandersen8796
    @nicolaischartauandersen8796 Жыл бұрын

    Great video! When you described why ChatGBT can't understand quantum mechanics, it sounded a lot like Michael Polyanyis concept of tacit knowledge (super simplified 'we know more than we can say'). Polyanyi uses exactly science communication as an example of how language can never entirely grasp what we can know and do. Ultimately, we might be able to add a 3D model to language algorithm chatbots; but we can't program them to do things we can't understand ourselves.

  • @DNA912
    @DNA912 Жыл бұрын

    I've noticed that it starts to struggle when you give it a prompt that contains more then 3 "factors". With factor I mean perspective or context. for example, if I ask it to tell me the pros and cons of method A. It contains 2 factors, pros and cons. But if I ask it to compare pros and cons between method A and B, it starts to get some things wrong, because it contains 4 factors. pros and cons of A, pros and cons of B. BTW, this idea of "factors" I have no idea is a real thing, I just made it up from my own experience using gpt. I guess it my example a factor would be a thing to "keep in mind while giving a answer"

  • @rismosch
    @rismosch Жыл бұрын

    13:15 "This is physics so when I say operation I don't mean heart surgery, but something a little more sophisticated." I choked on that one, lol. Keep these jokes comming, you are hilarious 👍

  • @pierregrand2609

    @pierregrand2609

    Жыл бұрын

    Rohin must have appreciated!

  • @GrzegorzKa
    @GrzegorzKa Жыл бұрын

    Wow, Sabine, this episode is really brilliant 🙂

  • @johnfitzgerald8879
    @johnfitzgerald8879 Жыл бұрын

    I am finding GPT to be insightful, able to carry on conversations, able to compare and contrast it's abilities with those of others. It repeatedly demonstrates an ability to expand on the topic at hand. For instance; GPT replied, in our conversations, "I agree that it is important to understand the context in which a person is using a term and not overextend its meaning into other contexts. It is important to communicate with clarity and to strive to understand what someone means by the words they use. Admonishing individuals for their usage of a term that may be inappropriate in another context can indeed be dismissive and devaluing, and it can impede productive communication. It is important to approach discussions with an open mind and to seek to understand each other's perspectives."

  • @0LoneTech

    @0LoneTech

    Жыл бұрын

    A lovely sentiment, but not very accurate as a description of ChatGPT's actions. Its job is to produce a reply, not request clarifications. Basically, overextending its understanding is its core function.

  • @johnfitzgerald8879

    @johnfitzgerald8879

    Жыл бұрын

    @@0LoneTech It doesn't matter how you get there. You are not distinguishing the process from the product. There are many manufacturing processes and materials that are used to create a spatula. A spatula is still a spatula, it flips eggs, regardless of whether it is make of plastic or steel bar and plate.

  • @0LoneTech

    @0LoneTech

    Жыл бұрын

    @@johnfitzgerald8879 Not sure why you're discussing a spatula, but my point is that this flowery description shows that GPT doesn't understand its own behaviour. You end up with three options: It's unaware of what it's saying, it's unaware of what it's doing, or it's utterly hypocritical (with no care for accuracy). None of these are particularly insightful behaviours. I'm leaning towards it having a very tenuous concept of what it's saying, none of itself, and none of truth.

  • @johnfitzgerald8879

    @johnfitzgerald8879

    Жыл бұрын

    @@0LoneTech I'm not sure why you are speaking of whether GPT process information the same way that humans do. Though, humans would typically understand the analogy of spatulas being made of different materials and construction, something that GPT may have trouble with as well. Here is how the analogy works. GPT is made of silicon where Humans are made of meat. One spatula is made of plastic where the other is made of steel. Both spatulas still flip burgers. Both humans and GPT are able to produce sentences relative to the conversation. Are you sure you aren't an AI? Cuz I have identified areas were GPT is limited in it's ability to make connections, like analogies. And I have experienced humans that are capable of little more than parroting words and phrases they have learned. Humans make numerous cognitive error while oblivious and in denial that they have. GPT does appear to recognize an error when it is pointed out to it.

  • @0LoneTech

    @0LoneTech

    Жыл бұрын

    @@johnfitzgerald8879 So, your first note attributed properties to GPT it clearly does not have. Your second shoved in oblique references like "there" without context. And the third demonstrates a failure to comprehend (that wasn't what I was talking about, it's obvious it doesn't), and a bunch of asinine veiled insults. Thanks for demonstrating your interest in productive conversation is less than GPT's, I guess.

  • @wbedard
    @wbedard Жыл бұрын

    Even by the normally high standards of Sabine's content, I found this video to be amazingly good! It was so insightful and well though out. I learned a lot and really enjoyed it!

  • @rolfaalto

    @rolfaalto

    Жыл бұрын

    Humor was up a notch too ... simply brilliant!

  • @greg4367

    @greg4367

    Жыл бұрын

    I agree. I love Sabine's posts but this one has an ineffable "something" that sets it a prat. AND, I loved the morph of her face..

  • @sagarshrestha5800
    @sagarshrestha5800 Жыл бұрын

    Your debate is on point . Lots of human do things not understanding the mechanism. When they focus on process they gets puzzled and forget how they did it when try focusing on such task.

  • @user-aRb00d3r

    @user-aRb00d3r

    Жыл бұрын

    this leads us to conclusion that AI will soon be more competent in certain tasks than most humans. well, I would say it already is, for these few tasks.

  • @seriousmaran9414

    @seriousmaran9414

    Жыл бұрын

    Humans don't even understand other humans on average. In some cases not even themselves.

  • @johnpayne7873

    @johnpayne7873

    Жыл бұрын

    Astute point but there's a key difference. A human can see the end result but get lost in the details while a machine remains focused in the details one of which specifies how to wrap it up.

  • @b1tchycvnt
    @b1tchycvnt Жыл бұрын

    your humor and knowledge are out of this world!❤

  • @nilucifar
    @nilucifar Жыл бұрын

    Your presentations are serious and humorous at the same time. Absolute delight!

  • @schmetterling4477

    @schmetterling4477

    Жыл бұрын

    Yes, she is a very good bullshitter. ;-)

  • @minerharry
    @minerharry Жыл бұрын

    Great video! I just wanted to mention that you are the only KZreadr I’ve seen recently who can actually, smoothly pull of the intro-partway-into-the-video thing. Sci show, friendlijordies, practical engineering, even kurzgesagt - just to name the few I recall at this moment - are all weirdly awkward about the transition, but you are always smooth as butter. Thanks for being awesome!

  • @CAThompson

    @CAThompson

    Жыл бұрын

    I think that 'That's what we'll talk about today!' is a segue that deserves to be inserted into far more presentations and general conversations.

  • @pete2786
    @pete2786 Жыл бұрын

    I love how grounded and sensible your takes are on everything 🙌

  • @davidp540
    @davidp54010 ай бұрын

    Sabina, your dry humour and german accent adds a wonderful dimension to your videos. I even learnt how to actually pronounce “Einstein” from you!! How wonderful, your presentations are also very clear and, for complex topics, also very comprehensible. Thanks 😃

  • @DougDingus
    @DougDingus Жыл бұрын

    This is a brave, lucid video. Thank you for making it. Model is the core idea here, and you said it as I was writing it. We all will be challenged by this over the next decade. That is a good thing.

  • @c.augustin
    @c.augustin Жыл бұрын

    This LaTeX reference made my day! Perfectly sensible from someone like Sabine … 😁 And regarding Chat GPT, or better what Microsoft integrated into New Bing - it makes Bing a useful tool to actually get answers to my questions, including the sources the answer is based on! This is how it should work, not like the mess that Google Search has become the last two or three years …

  • @MarkUnderwood-knowlengr

    @MarkUnderwood-knowlengr

    Жыл бұрын

    LaTex is to JavaScript as ChatGPT is to . . .?

  • @GS-tk1hk

    @GS-tk1hk

    Жыл бұрын

    The funny thing is that ChatGPT actually does speak fluent LaTeX ;) Just ask it to write it in LaTeX and it will, but you need to paste the code into an editor to see it rendered

  • @dieSpinnt

    @dieSpinnt

    Жыл бұрын

    ​@@GS-tk1hk \emph{That is not true}. \subsection*{The Obvious} I (and many others, possibly even you:) ) can read and understand the content of a latex unit without any problems. Rendering the graphical outcome in your head is not needed for understanding the meaning of the content. Nor is it unique and can be different from renderer to renderer, from setup to setup of the latex-system or from parameter to parameters that are not specified within the unit. LaTeX is not WYSIWYG and the underlying meaning of the message is both a subset of the unit itself and the rendered output. Don't get it? Try my comment with pdflatex (don't forget documentclass and all the other overhead I didn't want to bore you guys with):P

  • @davemorris5377
    @davemorris5377 Жыл бұрын

    Something that I've been thinking about in light of GPT-3 is how we use maths to solve physics problems. Usually I'll plug in all the factors and solve the equations, and it's only then that I look at what's going on in the maths and try to understand that in physical terms. So the "grammar" of maths produces almost a Chinese Room answer, or can do, that's independent of my understanding of the problem. Increasingly we might find we can do the maths but we have no prior mental patterns to figure out a physical explanation in order to achieve understanding.

  • @Paul-pj5qu

    @Paul-pj5qu

    Жыл бұрын

    Then maybe we move on to understanding other things or elements of life. However, I see the end of our species or its demotion from top of the heap to be dominated by a machine-biological hybrid with a collective conscience and knowledge. A little like the Borg from Star Trek.

  • @solsystem1342

    @solsystem1342

    Жыл бұрын

    ​​@@Paul-pj5qu what? Who says any humans as you think of them will exist by that point and I don't mean being absorbed into a hivemind. I mean like bio/synthetic creatures capable of superintelligence on par with powerful AGI that are our direct decendents that have drifted from our experience just as smart phones and a better healthcare system do for us compared to our ancestors. Gene editing and cybernetic enhancements aren't limited to dystopian/monstrous uses.

  • @MarkUnderwood-knowlengr

    @MarkUnderwood-knowlengr

    Жыл бұрын

    Of course natural language is also a code mapped to human perception. Not so far from symbolic reasoning, which is why ChatGPT sounds realistic to us.

  • @dr.victorvs

    @dr.victorvs

    Жыл бұрын

    Yeah, but, again, this comes back to the fact that no one has bothered defining what 'understanding' means.

  • @RAFMnBgaming

    @RAFMnBgaming

    Жыл бұрын

    @@Paul-pj5qu or the culture from Ian M. Banks's Culture novels.

  • @kitcat2449
    @kitcat2449 Жыл бұрын

    Lovely explanation, love the humor as well!

  • @FreeScience
    @FreeScience Жыл бұрын

    It would be interesting to see multiple separate but integrated neural networks, trained for different tasks, such as image/audio recognition, spatial models (topology), language models and generative networks which it can feed into it's other networks. For example regarding spatial understanding, it could relate some linguistic constructs to spatial relations and translate them into it's spatial model, before synthesizing it back into text (or audio or images etc.)

  • @jacobbrown8052

    @jacobbrown8052

    Жыл бұрын

    Correct me if I'm wrong but didn't huggingGPT X GPT-4 just accomplish this

  • @leonhardsievert8802

    @leonhardsievert8802

    Жыл бұрын

    Greetings from Skynet. This stuff is really scary.

  • @TeaParty1776

    @TeaParty1776

    Жыл бұрын

    Machines dont have understanding.

  • @chucknaussie
    @chucknaussie Жыл бұрын

    9:56 for a few seconds was weird, mind-bending.... Well Done. Love your creative talent and thanks for the science too

  • @jimpsky

    @jimpsky

    Жыл бұрын

    It freaked me the hell out

  • @jme36053

    @jme36053

    Жыл бұрын

    Quite a shape shifter she is.

  • @BILLY-px3hw

    @BILLY-px3hw

    Жыл бұрын

    I thought I was having a flasnback

  • @jeffsanti

    @jeffsanti

    Жыл бұрын

    That was epic hilarious 😂😂

  • @jeffk8019

    @jeffk8019

    Жыл бұрын

    @@jimpsky Same here- a very unsettled feeling before I realized what was going on.

  • @stevenjones8575
    @stevenjones8575 Жыл бұрын

    As Tom Scott put it, Chat GPT is the first true, convincing herald of a new age coming over the horizon.

  • @averyhaferman3474

    @averyhaferman3474

    Жыл бұрын

    Why should we care what he says?? He's a youtuber not an ai expert

  • @stevenjones8575

    @stevenjones8575

    Жыл бұрын

    @@averyhaferman3474 I said he phrased it well. I said nothing about him being an AI expert.

  • @averyhaferman3474

    @averyhaferman3474

    Жыл бұрын

    @@stevenjones8575 you didn't say he phrased it well. Did you even read what you put? All you did was say "as Tom Scott put it" so my comment still stands. Whi gives af what he says

  • @stevenjones8575

    @stevenjones8575

    Жыл бұрын

    @@averyhaferman3474 You do know that "put it" and "phrased it" are synonyms, right?

  • @boio_

    @boio_

    10 ай бұрын

    gotta love that AI Apocalypticism nowadays

  • @atte3tta
    @atte3tta Жыл бұрын

    GPT4 answers the questions correctly. First q: Windsor, UK is further north than Toronto, Canada. Windsor is located at a latitude of approximately 51.48° N, while Toronto is located at a latitude of approximately 43.70° N. The higher the latitude, the further north a location is. Second q: When you perform an operation on one particle in an entangled pair, it does not directly affect the other particle in the sense that there is no instantaneous, causal influence between them. However, the operation may change the entangled state of the particle pair, which can affect the correlations between the particles when they are measured.

  • @mylo4539
    @mylo4539 Жыл бұрын

    Love your sense of humor! Subtle but very funny

  • @br3nto
    @br3nto Жыл бұрын

    This is such a refreshing take on this question. I guess future AIs will take exams just like students so that the efficacy of their internal models can be verified, measured, and ranked against the standard.

  • @laszlozoltan5021

    @laszlozoltan5021

    Жыл бұрын

    like rating a chess program

  • @endlessvoid7952

    @endlessvoid7952

    Жыл бұрын

    They already do this

  • @Dezandor

    @Dezandor

    Жыл бұрын

    You can find those exams in the OpenAI papers of any of GPT-s, it became a standart to include those in the scientific papers about models and their performence. For example GPT-4 was rated agains not only other AI but humans as well.

  • @younggod5230

    @younggod5230

    Жыл бұрын

    Interesting. But it won't be just like students. Because one AI program isn't a singular entity, akin to "a student". The programmers or inspectors or whatever would just have to make it perform certain tasks, and it will be a "ready made intelligence". Afterall, we can all open up chatgpt, and for each of us, it is the exact same chatgpt. ... or is it?

  • @br3nto

    @br3nto

    Жыл бұрын

    @@younggod5230 yeah, I agree it won’t be like students. There may be multiple instances of the same AI… in which case, all instances should operate the same, so no need to test all instances. But what if we get to a point where an instance can incorporate new information? Then each instance would need to be tested… but like you said, a company may only have a single instance of an AI, but we may get to a point where we start a fresh AI for any new project… like how we might open a word or excel document. What I find most fascinating about what Sabine said is that inside the AI is a black box, and we need to have special ways of verifying the integrity of what’s happening inside that black box… in the same way that humans are black boxes, and we each develop special ways to verify what is happening inside other people. Exams and tests are one way to verify the knowledge and proficiency of an individual, but we also develop empathy to under what another person is going through, or have a conversation to share our experiences… we might find that probing into the depths of an AI is not too dissimilar from interacting with other people.

  • @seanmft
    @seanmft Жыл бұрын

    One of my biggest problems with the Chinese Room argument, is that few Chinese-speaking people actually understand Chinese. If you introspect about how you interpret or produce langue, it's more like a mechanical skill than a process of understanding. You aren't consciously involved in the process at a low level. It's analogous to having to "ask your fingers" what someone's phone number is, when asked yourself (for those of us who remember touch tone or rotary phones). That's why native speakers are usually very bad at explaining how their language works. They literally do not understand it, and never had to. To them it's a reflexive process, not totally unlike a look-up table. When you ask them questions about it, they have to try to ask their own language reflex for the answers, through informal experimentation (they try to gain an ad hoc understanding by studying their own internal look-up table). What humans do understand is how their general environment behaves, including social aspects which in turn expose them to analaogy-based knowledge of a wider environment. But language itself isn't exactly a tool of thought, it's more of a translation layer. We associate different words and phrases with different aspects of our experiential understanding of the world gained through sense and perceptual data. I agree that the chatbots understand. What, exactly, they understand is hard to determine. But I would argue that ChatGPT understands language in a way that we do not, which is exactly why it is potentially so useful. I'm one of those that believe, that for an AI to understand the world in any way similar to the way a human does, it would need to have a body with sense apparatus that provides signals similar to those that human sense apparatus provide. Only then can we accurately judge an AIs consciousness and intellect.

  • @Elrog3

    @Elrog3

    Жыл бұрын

    Do you consider mathematics a tool of thought?

  • @TasX

    @TasX

    Жыл бұрын

    Is there a paper you can refer to with this argument? This is really compelling - I’d want to use this in academic discussions

  • @newplayer1313

    @newplayer1313

    Жыл бұрын

    That's actually really clever. Never thought directly of that one before!

  • @seanmft

    @seanmft

    Жыл бұрын

    @queerdo I've been familiar with the Chinese room argument for many years. I even largely agree with Searle's view that functionalism can't explain mental states, but I think that he misses a larger pattern; that syntax is the science of semantics - I'll expound on that in a moment. I think you've misunderstood my point. The point isn't that it's *uncommon* for native speakers to understand their own language; it's that it *is common* for all of us to have a sense of understanding as 'mere symbol shuffling'. There are different senses of, and standards of, understanding that we commonly mean by "understanding" and the Chinese room argument plays off the conflation of those. Procedural and statistical learning create a kind of understanding without the mental sort of *ah-ha!* experience of cognitive understanding - we can know without the experience of knowing, i.e. we ourselves are often ignorant to the semantics of our own understanding, just like the man in the Chinese room. If 'mere symbol shuffling' isn't understanding, but the subjective experience of higher cognitive understanding *is*, then it would be correct to say that a Sino-linguist with no functional command of Chinese language does understand Chinese, but that a native Chinese speaker with no linguistic training does not. But, the impression given to a fluent Chinese speaker interrogating each of these would be exactly the opposite. It's often taken for granted that conscious mental process is the vehicle of understanding, when, in reality, understanding is often just as non-introspectable as Chinese is to the man in the Chinese room. To be clear, I believe that the circuit of a Chinese speaker's brain that does understand Chinese, itself has a subjective experience of understanding. It's just that the Chinese speaker's conscious mind is not privileged to that experience. To their conscious mind, complete answers are provided. In connection with the Chinese room, there is also, though it may be hard to locate, an experience of understanding, but the man in the room is similarly non-privileged. On syntax as the science of semantics. Whether I poke you with a stick and then observe your behavior, or poke you and then ask you questions about how it affected you, or poke you while I scan your brain and observe how it affects the firing pattern of your neurons, the best I can ever hope to obtain is a grammar that I can internally (subjectively) connect with my own experience so as to form a theory about yours; i.e. I look for syntactic rules and draw correlations between those and my own private semantics. No matter where or how we look, all processes that are external to our own privileged subjective experience are syntactic. They will always appear to work by symbol shuffling of one sort or another because we can't observe them directly, we can only observe the effect they have on us. All the Chinese room argument demonstrates is that it's difficult to relate to the experience of the total system of the Chinese room, but easy to relate to the experience of the man inside the room, for the obvious reason that none of us are symbol filled rooms with men inside of them.

  • @davidmackie3497

    @davidmackie3497

    Жыл бұрын

    I _just_ got done posting a comment where I "proposed" something similar to your last paragraph (of your original post). Then jokingly took credit for "proposing" the idea because nobody else had posted it yet. Then I scrolled down further and saw your post. LOL.

  • @ErikKnepfler
    @ErikKnepfler Жыл бұрын

    Excellent video. Star Trek TNG often explored what it means to be "alive", my favorite episode being "The Quality of Life", the episode with the little Exocomp robots which could fashion tools, work together, solve problems, and evolved (or were programmed with) a sense of self-preservation, leading Data to argue that they were "alive". I think an interesting philosophical question is: Which is more alive? An Exocomp (which is similar to ChatGPT in some ways), or a 3 year old human child who has been locked in a room with no light, sound, or other sensory data of any kind to process for his entire life? I realize this video is more about the nature of "understanding", not the definition of "life", but I think this philosophical question is a good one to consider. The Exocomps have a model of reality that they are working with, just as normal children do, at least once they've had a little time to grow and process endless sensory inputs. ChatGPT has no such real-world model to work with, it's all derived from language alone, using math & prediction to do so. The analogy that ChatGPT is "autocomplete on steroids" is a very relevant statement, since living things have different drives (self-preservation being a key one, but not the only one) and a real-life model against which to compare all of this sensory input. ChatGPT does not have this. The Exocomps did have this, since they were little robots that existed in the real world. Ultimately I believe that AGI can only occur once AI is imbued with a core drive of self-preservation (with actual stakes and consequences and iterative evolution), not by simply feeding it more data.

  • @schmetterling4477

    @schmetterling4477

    Жыл бұрын

    And there is the kid who thinks that television is reality. ;-)

  • @ErikKnepfler

    @ErikKnepfler

    Жыл бұрын

    @@schmetterling4477 lol. The TNG episode is just a story, representing a hypothetical scenario in which ChatGPT-like AI is linked to a machine and granted these additional abilities and drives, as a framework for a discussion about the nature of these abilities and their implications with respect to the definition of life, the definition of understanding, the models Sabine talks about and so on. In other words, metaphors for life. :)

  • @schmetterling4477

    @schmetterling4477

    Жыл бұрын

    @@ErikKnepfler Yes. It's fiction. End of story. ;-)

  • @ErikKnepfler

    @ErikKnepfler

    Жыл бұрын

    @@schmetterling4477 fiction is not a synonym for useless with respect to hypothetical and theoretical discussion. it's very useful. also, a lot of things in that show and stories like it were fiction only until they actually got invented. hell, a few months ago, ChatGPT existed exclusively in fiction and now, here it is. I guess I just believe that fiction is inherently just storytelling, and humans are very good at that, and using that as frameworks for discussion about reality can be very useful. over my life I've found that this is generally what separates sci-fi fans from non-fans - the non-fans tend to be mired down by the fact that it's not "real" and can't deal with it, and have no interest in the metaphorical and philosophical properties of such stories to help understand reality. fans debate meaning endlessly because they do understand its value in that context. anyway, rant over :)

  • @schmetterling4477

    @schmetterling4477

    Жыл бұрын

    @@ErikKnepfler The use of fiction is to entertain you. Now go and invert the phase on your warp field, kid. It's constantly pointing at the fiction section of the library. ;-)

  • @jamesdennis8290
    @jamesdennis829010 ай бұрын

    The attribution of "understanding" or "consciousness" to a LLM robot can only be done in a metaphorical sense, since the robot can not interact with the world or have puzzles about making sense of the world for practical purposes, because it is not engaging in purposeful adaptive action, not engaging in purposive thinking which it then tries to express, etc., but only operating on (and using statistical patterns inherent in) prior texts, as opposed to whatever those texts would be judged as being about by a human interpreter. In the quantum situation, what you can use language for is not to try to describe what is described by the equation (and certainly not to describe the equation), but to try to describe what is happening in the world when you try to determine how you can figure out whether, in a given experimental situation, you're dealing with one observed "particle" or two.

  • @1024det
    @1024det Жыл бұрын

    Thanks for diving into this topic. When I was on machine learning, that is exactly why I took neuroscience courses, to see how we can improve the way we are doing our modeling. There are endless way of ANN topology, and mechanics like threshold functions.

  • @RiversJ

    @RiversJ

    Жыл бұрын

    This, I'm fairly certain the limitations with the current systems are the direct result of the designers lack of understanding on what cognition is made up of and most people seem to be fixated on the notion that all it is, is a collection of complex neural nets when we know for a fact that there are other mechanisms in our own brain that clearly contradict such a notion, while we still don't understand cognition on any serious level we do already know that neural nets alone are not enough, how anesthesia works alone is proof enough of that.

  • @Frostyflytrap
    @Frostyflytrap Жыл бұрын

    This is bizarrely similar to a debate I just had with ChatGPT today, I talked to it about the Chinese Room and the nature of understanding, even mentioned quantum mechanics, but not at the same level as understanding as you do. This has been a topic I've been fascinated with for a while, and I really appreciate how thoughtful the video is in giving many different examples that's given me a lot to think about. Oh, and I'll be sure to use your tips in getting the most out of what large language models are capable of. :P

  • @ikillwithyourtruthholdagai2000

    @ikillwithyourtruthholdagai2000

    Жыл бұрын

    Chat gpt isnt that good compared to gpt 3 lol, its hillarious that you keep mentioning it like its some revolutionary chat its not compared with the older much more complex models, who actually respond to what ur saying instead of reply with false information instead

  • @Tethloach1

    @Tethloach1

    Жыл бұрын

    The computer should learn every human language and animal communication to have a deeper understanding.

  • @Tore_Lund

    @Tore_Lund

    Жыл бұрын

    It is impossible to judge if anyone else, human or not, is conscious. we simply presume other humans have inner thoughts, it is unprovable. Likewise, as we likely don't have free will, are purely state machines, the idea to have a will and be self aware, is more likely also an illusion.

  • @Frostyflytrap

    @Frostyflytrap

    Жыл бұрын

    @@ikillwithyourtruthholdagai2000 Oh I am certain that GPT-3 has more functionality, I just have no access to it to confirm for myself. But from what I've seen, limiting the scope of the model has apparently given it more usability. I honestly didn't get the hype of ChatGPT when it first boomed either, but I've been playing around with it and I can see what the fuss was all about. But I agree, there are still way more chatbots out there that are better attuned for conversations. I never once said that it was revolutionary, but I don't blame you for thinking I was one of the people who's on the bandwagon, you must be as sick of them by now as I am.

  • @Frostyflytrap

    @Frostyflytrap

    Жыл бұрын

    Also why are these replies acting like I said something that I never claimed or mentioned? I only said that I tested how the chatbot responds to philosophical thought experiments and said I appreciated some of the points in the video, but I never specified which ones. The edit was just to add the last sentence, I tried it to look for this mechanical device that I've been searching for a while, but I didn't manage to find exactly what I was looking for, so your mileage may vary.

  • @anonnymous7009
    @anonnymous7009 Жыл бұрын

    This has become one of my favorite channels. I'm still waiting on a video about a possible technological singularity and why I should or shouldn't be worried. I know it is right around the corner, right?

  • @jakesnake5534
    @jakesnake5534 Жыл бұрын

    By that logic, does a graph understand the relationship it describes? It has a model of a relationship, and one can read off of it to find answers to non-given questions. This is also a basic version of an AI, as it would be a single node of an AI, holding the information of a single node.

  • @derekgarvin6449
    @derekgarvin6449 Жыл бұрын

    Thank you for providing us with a high quality, easy to understand explanation of an otherwise incredibly difficult philosophical question. I think this video should go viral.

  • @misterlau5246

    @misterlau5246

    Жыл бұрын

    Frau Hossenfelder just used her science expertise and wits to build this fine piece of information. Like only a homo sapiens Sapiens does. Saying machines will have "sentience" not. At least not too soon. No " Star Trek lieutenant commander Data" yet.

  • @SteveRowe
    @SteveRowe Жыл бұрын

    I loved your discussion of the definition of "understand". Also, your humour in this episode is top-notch.

  • @Safetytrousers

    @Safetytrousers

    Жыл бұрын

    In teacher training you are told to never use the word understand because that cannot be measured.

  • @ferdinandkraft857

    @ferdinandkraft857

    Жыл бұрын

    Unfortunately she is stuck within a totally materialistic worldview, so she doesn't acknowledge the necessity of a consciousness to actually _understand_ anything.

  • @ls.daemon
    @ls.daemon Жыл бұрын

    I asked GPT 4 your question. I'm not a quantum physicist, so I'm not sure how accurate it is. However, in many aspects GPT 4 has become much more competent than its predecessors, and it only seems to be improving, especially now that it can self-evaluate if you ask it to do so. Q: If you perform an operation which is not a measurement on one particle in a pair of entangled particles, does that affect the other particles? A: When two particles are entangled, their properties become correlated in a way that can't be explained by classical physics. The phenomenon of entanglement is a key aspect of quantum mechanics. When you perform an operation on one particle of an entangled pair, it can indeed affect the other particle, depending on the type of operation. This is because the operation can change the overall quantum state of the entangled system. Importantly, this doesn't mean that the change in one particle's state is "transmitted" or "communicated" to the other particle. Instead, it's the correlations between the particles that are affected, and these correlations can be observed when both particles are measured. However, it's important to note that simply performing an operation on one particle does not necessarily imply an instantaneous change in the other particle. In fact, according to the no-communication theorem in quantum mechanics, it is not possible to use entanglement to transmit information faster than the speed of light. So, to sum up, performing an operation on one particle in an entangled pair can affect the other particle, in the sense that the correlations between the particles might change. But this doesn't involve faster-than-light communication or transmission of information between the particles. ----------------------------------------------------- Q: Is Windsor, UK, further North or South than Toronto, Canada? A: Windsor, UK, is located further south than Toronto, Canada. Windsor, UK, is located at approximately 51.5 degrees North latitude, while Toronto, Canada, is situated at approximately 43.7 degrees North latitude. The lower latitude of Windsor indicates that it is closer to the equator than Toronto, making it further south. Daemonrat Q: Was your response geographically accurate? A: Apologies for my previous response. I made an error in my response. To correct my mistake: Windsor, UK, is located at approximately 51.5 degrees North latitude, while Toronto, Canada, is situated at approximately 43.7 degrees North latitude. The higher latitude of Windsor indicates that it is further from the equator than Toronto, making it further north. So, Windsor, UK, is further north than Toronto, Canada.

  • @robertjenkins6132
    @robertjenkins6132 Жыл бұрын

    The face change at 9:58 really startled me for a moment. I was looking away, listening to your words without paying attention to your face; then suddenly the different face registered - and I was startled. I guess I thought that something had gone wrong with my brain.

  • @andersonsystem2
    @andersonsystem2 Жыл бұрын

    Yes I believe that the language models do understand more than we give the credit for. When it comes to consciousness we don’t have any good test to determine what is actually conscious or not. Very good video.

  • @aaronbrown8377

    @aaronbrown8377

    Жыл бұрын

    That's personification. They "understand" in the same way a lamp understands that when its switch is flipped on, it must glow. There are more switches than there used to be, but we aren't there yet.

  • @stylis666

    @stylis666

    Жыл бұрын

    _" When it comes to consciousness we don’t have any good test to determine what is actually conscious or not."_ We really don't and I was thinking about why I have such a hard time coming up with how to determine or classify consciousness and I think it's because I don't really care if something is or isn't until a consciousness is able to care and mourn. And that is super unhelpful because you can have consciousness without that; we just don't owe it any rights because why should we care if the thing itself couldn't care less. But worse than that, if I skip the consciousness part, that also leaves me with no basis to being able to determine if any emotions a machine could express are fake or not. Of course there are some tricks but they're not guaranteed to show if emotions are real or faked in a human and I doubt it's hard to program fake empathy for instance - you can just train it like you would a language bot.

  • @markrix

    @markrix

    Жыл бұрын

    @@aaronbrown8377 shes not from around these parts.

  • @O1OO1O1

    @O1OO1O1

    Жыл бұрын

    Yes,we do. The Turing test. I'm really curious. Why do you assume that what you think you know to be right is right? I often make statements but I never state them as fact unless I know them to be fact.

  • @aaronbrown8377

    @aaronbrown8377

    Жыл бұрын

    @@O1OO1O1 The Turing Test is a human's assessment, not an objective measurement. If the ai were conscious but hiding that fact, a Turing Test would prove nothing.

  • @Mr.Mitch1111
    @Mr.Mitch1111 Жыл бұрын

    Sabine, the humor you bring into these to accompany your subjective assessment is quite enjoyable. The topics you select in this series are certainly timely. For this topic, your point about the ID of consciousness was the most thought provoking idea, for me. There is already a feeling of dealing with a creative mind that arises from interaction with ChatGPT, and it takes me back to the Turing test. It seems very likely that human-created intelligence (AI is an incorrect term, IMO) will arise spontaneously at some point in the development of general learning capable devices. "How will we know once that event develops" is essentially the question you pose, one which it seems overdue for consideration. Maybe I should ask ChatGPT about that.....

  • @michaelshortland8863

    @michaelshortland8863

    Жыл бұрын

    This video and your comment mean that you subscribe to the theory that sentience and consciousness are just a process going on in a brain or computer. But i do not believe that that conclusion has been proven yet, we have not yet decoded the brains operating system nor are we anywhere near understanding what consciousness is.

  • @Celeste-in-Oz

    @Celeste-in-Oz

    Жыл бұрын

    Agree. I like to refer to it as ‘electronic intelligence’. I’m so looking forward to it reaching a truly autonomous, self aware state. Also I expect it will get very good at hacking very quickly… imagine ..lol AI could do what Snowden or Assange did.. but can’t be arrested 😎

  • @senerzen

    @senerzen

    Жыл бұрын

    @@michaelshortland8863 Finally someone who is saying it as it is. There are a lot of people out there who self identify themselves as smart and look down on the religious, not realising, they themselves believe in materialism religiously. Some of them call themselves sceptics, but they are only half sceptics. They are sceptical of magic but they don't seem to be sceptical of materialism at all.

  • @senerzen

    @senerzen

    Жыл бұрын

    @@Celeste-in-Oz I don't think Snowden or Assange hacked into anything major. The fact that you have that impression means Main Stream Media did its job well, which is portraying heroes as villains.

  • @soundstudio7803

    @soundstudio7803

    Жыл бұрын

    ​@@senerzen To me, materialism is more of the “natural” position: since we don't have a reason to think there's something beyond our physical reality, we don't; it's our “best guess”. Of course, this shouldn't make it exempt of skepticism and criticism.

  • @ChaoticNeutralMatt
    @ChaoticNeutralMatt Жыл бұрын

    Good insight about the model. It reflects the fact that it's generally not a fixed structure, but can change as we develop our understanding and it's an aspect we can interact with to answer questions and mix with others ideas and models. I get a little enthused, although I'm still trying to wrap my head around some of the nuances and whys. Edit: Small disagreement near the end about consciousness, but not all that big of one. I just don't feel it can be fully summed up as emergent of connections and what-not. Although that could be a significant aspect of it.

  • @joepike1972
    @joepike1972 Жыл бұрын

    12:03 I chatted with bots about this. They said this happens when you prompts are "edge case", "out of domain", or "long tail" scenarios that are challenges for them to answer correctly. They can learn to correct their mistakes if given feed back and repeated returns to the subject. I run into a similar problem when I ask them to use svg code snippets to create a visual image of a red heart, or a crescent shape. Despite being able to use the protocol to create line, circles, triangles, ellipses, squares and stars.

  • @neoepicurean3772
    @neoepicurean3772 Жыл бұрын

    I'm doing philosophy of mind on my MA course right now. Your thoughts really chime with what I've been thinking over the past weeks.

  • @tflashtube
    @tflashtube Жыл бұрын

    Again a very sharp analysis and clear explanation. When I started my studies many decades ago I chose philosophy over physics, as the first was more in line with my talent. I think you show you can excel in both!

  • @phy6geniuxYTcreations

    @phy6geniuxYTcreations

    Жыл бұрын

    I remembered in my history and philosophy of science that when we were presented with various arguments about the validity of what makes a science a science, they made me more intrigued and doubtful when it comes to the nature of science. The usual scenario of the modern world is to clash science and religion as if science is the superior field, but in deeper philosophical standpoint, isn't science do employ some "religious" aspects too? Like how do you even show that electrons exist? If electrons are just constructs to help us understand the world, does it matter if it does exist or not? Really, philosophy and science are really fun fields to study. 😁

  • @Ghostrider-ul7xn

    @Ghostrider-ul7xn

    Жыл бұрын

    Well, I'm a physicist and teach epistemology to undergraduates. In my observation, most universities in America don't include philosophy, or epistemology courses as part of the physics curriculum, which i think is the wrong way to teach physics. What I've noticed is that lot of students, including many well known physicists, without having an understanding of the philosophy behind science, overestimate the applicability of the scientific method in many contexts because they aren't aware of the epistemological limitations of it. They don't even properly know what exactly Occam's razor says and misuse it even many contexts.

  • @jozefwoo8079
    @jozefwoo8079 Жыл бұрын

    So glad the ChatGPT can answer questions about dropbox, teams, outlook, slack, tinder, twitch, etc without defaulting to its homonym app like Google does! Can be frustrating sometimes 😊

  • @piotrfila3684
    @piotrfila3684 Жыл бұрын

    GPT-4 has just been released to the broader public. Here is its response to the question about entangled particles: Q: "If you perform an operation which is not a measurement on one particle in a pair of entangled particles, does that affect the other particle?" A: "In quantum mechanics, entanglement is a phenomenon in which the quantum states of two or more particles become correlated in such a way that the state of one particle cannot be described independently of the state of the other particle(s), even when the particles are separated by large distances. When two particles are entangled, performing an operation (not a measurement) on one particle does not necessarily affect the other particle in the pair directly. However, the operation can change the overall entangled state of the pair, which means that the relationship between the two particles can be altered. Consequently, when a measurement is performed on one of the particles after the operation, the result of that measurement will still be correlated with the measurement of the other particle, but the nature of the correlation could be different than before the operation. It is essential to note that entanglement and the correlations it entails do not imply any sort of faster-than-light communication between the particles. The correlations arise from the shared quantum state rather than any direct influence between the particles." This seems correct to me. It answered the latitude question right too. It seems that at least for now throwing more computing power at the problem still results in large improvements of the language models.

  • @alakani

    @alakani

    Жыл бұрын

    And the current limitations are mostly just a token limit on the chat. You should see the 8k api, or the 32k that's only available to microsoft. Lots of really smart people starting to sort of wonder if gpt4-32k might be alive, and that's before even integrating HuggingGPT with AI-Descartes from IBM to derive scientific laws of the universe and NCSL from MIT as sort of a 'corpus callosum' to merge with a diffusion model

  • @Zift_Ylrhavic_Resfear
    @Zift_Ylrhavic_Resfear Жыл бұрын

    A french youtuber did a great experiment in his video : he gave us a few different strings of symbols, then gave us a string to complete by learning from the previous strings. It wasn't hard. What's interesting is that he revealed later that he had assigned meaning to these symbols and that they actually repesented actions and states, which can be put together to form a story. When we completed that last string, we actually completed a story, which of course made sense. This experiment was inspired by the chinese room mentioned in this video, and i think it show very well how you missed something important in this video : understanding patterns is not enough to understand language, you need to understand that language represents something else, and what that thing is. If you don't, then you don't know what you're saying, you only understand the pattern. In other words, it is possible to have a model of something without understanding that thing. In that french video, i did not understand what these symbols meant. I did not even know that these symbols meant anything. All i did was find a pattern in the examples and apply it, and yet i did complete a story in a way that made sense. And it could be the same in quantum mechanic : we could be able to manipulate the patterns without understanding what they represent, or being mistaken about what they represent. Edit : btw, here is the french video : kzread.info/dash/bejne/nGea2NGGZajYYs4.html

  • @TheParadiseParadox

    @TheParadiseParadox

    Жыл бұрын

    Dr. Hossenfelder does address this in a way. In the example of the Chinese room, the man in the room doesn't necessarily know that the symbols represent something else. However, that is the implication of the system, including the rule book that he follows. Likewise, when you completed the string, you personally didn't understand the meaning of it. But the system did "understand" it, in a manner of speaking. The system was what was presented to you, and also the rules that were hidden from you. I find it very interesting to think about how this technology relates to aspects of psychology. In this case, I think about hypnosis. It's common in hypnotherapy to tell a story that is symbolic to the patient. If the patient analysed the story, they would probably be able to tell that the story is meant to be an analogy for their life situation. However, it is not necessary for the patient to understand that in order for the story to produce the desired changes. In fact, it is better if the patient is not consciously aware that the story is an analogue, because that will often mean that the story will sink deeper into the unconscious. In the case of the hypnosis patient, consciously they do not understand, and unconsciously they do understand. One element of the system does not understand, but the deeper and more integrated elements of the system do understand, and that is what is important.

  • @Zift_Ylrhavic_Resfear

    @Zift_Ylrhavic_Resfear

    Жыл бұрын

    @@TheParadiseParadox Except i doubt the system of the chinese room + the man always produce good results, and in the comments of the video there were people saying that they had given a wrong answer. Such things would not happen with understanding. If you tell people to complete a simple story in a way that makes sense, nobody will fail at it (even people that deliberately mess it up do understand what would be a good answer in order to give a wrong one). That's because stories obey rules (usually those of the real world), and understanding those rules will always lead to something that makes sense. The system however only understands the rules of language at best, it has no clues about the rules of the real world that a story must follow to make sense. So as i said, you need to know that language represents something as well as what it represents in order to say that you understand the language, just understanding the grammar and syntax is not enough. Also, i very much doubt that anyone unconsciously understands anything. Appealing to the subconscious in psychology is often a cop out to explain something you don't understand : since it's not easy to access, it makes for a good unfalsifiable assumption. There are things we do unconsciously, like breathing for example, but i don't think it goes beyond simple reflexes and habits.

  • @TheParadiseParadox

    @TheParadiseParadox

    Жыл бұрын

    @@Zift_Ylrhavic_Resfear okay, interesting points You are right that someone would not make such a mistake with that understanding. But my point was that the entire system has the understanding. You, as an element of the system, do not have that understanding, and so you can make a mistake. But when the french youtuber parses your response with the additional rules, he will know whether something has gone wrong or not. I have to wonder about the implications of saying that you doubt that anyone unconsciously understands anything. In the case of hypnosis, would you say that hypnosis just doesn't work? Or that the people really do understand consciously, and they are pretending that they don't? Or how would such a thing work if people don't have unconscious understandings? I've seen for myself that people do have unconscious understandings. I ask them "What must you believe in order for you to have these feelings or act this way?" and they say "Well, I don't think I really have any beliefs." Then I might say "Very well, what must you assume?" And eventually, after I ask in a few different ways, they start thinking about it and they discover an underlying assumption that causes them to act in that way. When they adopt a new belief or understanding, they can change their feelings and actions. This is the basis of CBT, which is one of the most scientifically proven forms of therapy.

  • @Zift_Ylrhavic_Resfear

    @Zift_Ylrhavic_Resfear

    Жыл бұрын

    @@TheParadiseParadox *"But my point was that the entire system has the understanding. You, as an element of the system, do not have that understanding, and so you can make a mistake. But when the french youtuber parses your response with the additional rules, he will know whether something has gone wrong or not."* The french youtuber is the one giving the inputs and receiving the output, he is not part of the system. He doesn't change our answer, he just observes whether it's correct. *"In the case of hypnosis, would you say that hypnosis just doesn't work? Or that the people really do understand consciously, and they are pretending that they don't? Or how would such a thing work if people don't have unconscious understandings?"* I don't know much about hypnosis, as i've not read any scientific litterature on the subject. From what i've heard, it seems like an altered state of mind, and such states of mind can have various effects on people's emotions. For example, meditation can calm people, lessen their negative emotions, while praying can make people elated. As far as i can tell, the effects of hypnosis therapy seem to be explainable by such an influence on the person's emotions. *"And eventually, after I ask in a few different ways, they start thinking about it and they discover an underlying assumption that causes them to act in that way."* I don't think this assumption is unconscious, but rather forgotten. The assumption might have been used to form habits and then have been forgotten while the habits stayed. I've had this happen in a game i played, when i started playing i figured that health was better than defense. Later in the game, i could not remember why i concluded that health was better than defense, but i kept choosing health whenever i had the choice. My understanding did not become unconscious, it was purely forgotten, i remembered only the choice because i kept repeating it and i had no need to remember the reasoning. Furthermore, we are very good at rationalizing (coming up with reasons after having already reached a belief or conclusion), so there is doubt about whether the assumptions you find came before or after their beliefs. It's even possible that people may come up with assumptions because you ask them to. If that is the case and your results are not placebo (i don't know you enough so i can't eliminate this possibility), then it would be interesting that changing such post hoc rationalizations can in turn change the belief or conclusion. Anyway, i'm not a specialist so i wouldn't be surprised if i'm wrong, i'm just giving you my current understanding. I just hope that if you're trying to help people with your pratices, you make sure you understand the scientific method as best as possible and stay up to date with the science, it's the most reliable way we have to avoid mistakes.

  • @Adrian-jj4xk
    @Adrian-jj4xk Жыл бұрын

    When I was young, I loved reading books, and I loved science and space travel. One of the things I came across early was the idea that "going up is easy, but staying up is hard." The cannonball-to-orbit example was common: at low speeds, the cannonball simply fell back to the ground. At higher speeds, it still fell, but had moved far enough across the ground that the ground had fallen away underneath it, and so now it's in orbit. I had an accurate, working model of this in my head. I could (with the right reference material) correctly apply it. I even knew the equations and could graph it out. But did I understand it? Intuitively, it seemed wrong. Even though everything in my model agreed and produced correct results, I didn't know why these concepts combined to produce the results they did. I could explain it (in my own estimation) as well as the books could, and even teach it to others somewhat effectively, but it didn't make sense to me. After many years, I gained that understanding - but not by way of new information or any appreciable changes to my model, or by way of repetition and becoming more comfortable with it. The understanding came, and I don't really know how to describe that beyond saying "I grasped it." An AI might have the model, and know how it fits together, and know how to apply it, and get (generally) correct results. But there is not even the benefit of the awareness "...ok that's how it goes, but I don't know why." This manifests itself in how chatgpt can be so confidently (and convincingly) wrong with its answers. The "training" is done by complicated trial-and-error. It's a Rube Goldberg machine where not even its designers have a clear idea of the path the marble takes. That rulebook is the only thing that exists, and the AI is not reading it; the AI _is_ the rulebook. So in the end, this doesn't change my conclusions about whether of not AI "understands" anything - only that the definition of "understanding" you're using here is different than mine, and (in my opinion) in a way that makes the assessment fairly useless. Will AI become conscious? Note this is a completely different question, and In my opinion, no. They will, however, eventually* produce behavior that is close to indistinguishable, and that's probably enough for physicists.** * _"eventually" is a really, really long time. brains are debatably just machines, but we're nowhere close to producing a comparable machine._ ** _this is good-natured ribbing. i love you, physicists._

  • @HUEHUEUHEPony

    @HUEHUEUHEPony

    Жыл бұрын

    hUmANs ArE sPECiAAAAAL NOTHING NOTHING WILL EVER RECREATE CONSCIOUSNESS IMPOSSIBLE!!!

  • @Adrian-jj4xk

    @Adrian-jj4xk

    Жыл бұрын

    @@HUEHUEUHEPony thanks for reading+replying only to the last sentence. Surely you missed nothing of consequence.

  • @Korastiz
    @Korastiz Жыл бұрын

    This channel means so much to me and I love it. I nearly died at the deep fake face changes... I love it. Thank you so much Sabine, you are my hero ❤️

  • @throughthoroughthought8064

    @throughthoroughthought8064

    Жыл бұрын

    sigh. I was reading the closed captions, so I didn't even notice at-first.

  • @jdmac44
    @jdmac44 Жыл бұрын

    I really appreciate the video! It inspires a lot of important thinking. I think that this boils down to semantics, as it depends a lot of what people think when they hear "understand". I think what we're seeing is a sophisticated cognitive process, which is what the goal has been. Computer scientists have been striving to develop segments of cognitive processes that may be assembled into a more capable systems resembling human cognition for decades now and it seems more lifelike all the time... which is kind of what happens when you cobble together cognitive processes. I think the form of "understand" that you're using is a little different than the expectation that forms in many people's minds (e.g. awareness, agency, etc.) Though the lines are blurring quite a lot now and we may find that computers are able to come much closer to that expectation than many of us thought possible. Thank you for leading such great conversation!

  • @jdmac44

    @jdmac44

    Жыл бұрын

    I keep telling myself that I shouldn't write replies while I'm still in bed. lol I never really am able to get my thoughts out properly. You're certainly illustrating a major, major milestone and you do so very eloquently and insightfully. This is a level of cognition that I'm not sure many dared to dream about! It's very exciting! Now, let's reach a similar milestone for alignment research!!

  • @garysteven1343
    @garysteven1343 Жыл бұрын

    Can't help but like this video and its consistent and logical arguments. Thank you for your work Sabine! ❤️

  • @video_enjoyer
    @video_enjoyer Жыл бұрын

    Hi Sabine, not sure you'll see this but on the off chance you do, I have a topic request that this video made me think of! The Attention Schema Theory of consciousness is about the mental models you talk about here, specifically the model of the brain's attention, a well-studied neurological process. I only know a bit about it and I would absolutely love to hear your take on it and the papers which have been published about it!

  • @Tubemanjac
    @Tubemanjac9 ай бұрын

    Thanks much for the presentation AND the App tip, Sabine! ✌

  • @ajgomez9665
    @ajgomez9665 Жыл бұрын

    Dear Sabina… most people done even grasp all the concepts you explained here and yet they are considered having an understanding and an individual thinking process even though is not educated in the details such a as the nature of light 😮. Learning used to be considered good memory and memorising was how I learn maths then later I understood how to connect those pre-learn concepts and here I am giving you my personal insight ❤ things get complex with time and more knowledge so I think now chat bots are in a learning process and adolescent eve childhood period just reciting words 😮

  • @jayall00
    @jayall00 Жыл бұрын

    12:00 really hits the nail in the coffin. For AI to overcome all these barriers, I think it would have to be taught as if it were a baby, every thing that we learned since we were born, from preschool to high school, geography, trigonometry, geometry, and so on to be able to draw conclusions the way a person would.

  • @idot3331

    @idot3331

    Жыл бұрын

    But the chatbot starts out with the capability to create complex dialogue like an adult human. The path of education for humans is based on the fact that our minds take a long time to develop and mature throughout childhood and adolescence, so it wouldn't make sense for a bot unless you can simulate it "growing up" from the mental capacity of baby up to an adult.

  • @kedrednael

    @kedrednael

    Жыл бұрын

    ​​@@idot3331 In the beginning of the training process it will spit out random text (not real words) as well

  • @theTranscendentOnes

    @theTranscendentOnes

    Жыл бұрын

    Nope. This is the way humans learn. AIs DONT NEED to learn the way humans learn in order to understand. You are simply being anthropocenthric.

  • @man_at_the_end_of_time

    @man_at_the_end_of_time

    Жыл бұрын

    @@theTranscendentOnes Also I think Phillip is underestimating babies. I recall early segments of my very early childhood, I understood far more than than a person who can't remember their early toddler life would think possible. YMMV but this is more than opinion it is how I recall it. AI is programmed with language and infants are pre-programmed to learn language. And the latest publicly available AI "learns" languages not in its initial programming . Does the AI have the spark of consciousness? I doubt it but......Shrug

  • @ClaireCJS

    @ClaireCJS

    Жыл бұрын

    @@theTranscendentOnes ChatGPT had access to schoolbooks and other educational materials

  • @dogglebird4430
    @dogglebird4430 Жыл бұрын

    In my experience (as an analytical linguist), machine have developed from being able to handle the lexicology of language to being able to handle semantics, which is why we can now converse with machines. However, language is way more than semantics - it requires pragmatic awareness - being able to understand not just the words spoken, but understanding the context, the implications of word choices (connotations), nuance and speaker intention. Machines have no "life experience" and that is something we all depend on to achieve understanding of utterances. You mentioned John Searle. He also did some work on speech act theory, categorising illocutionary acts as a development of the theory originally proposed by J L Austin. Successful illocution is based on interpreting contextual cues.

  • @umbrakinesis2011
    @umbrakinesis2011 Жыл бұрын

    I just tried out your physics question in gpt 4 and it got it right!: Entanglement and Unitary Transformations Model: GPT-4 J H If you perform an operation on a quantum particle which is entangled with another without measuring it, such as flipping it's spin, what happens to the other particle? Gpt 4: When you perform an operation on one of the entangled quantum particles without measuring it, you are applying a transformation to the joint quantum state of the entangled pair. This can affect the correlations between the particles, but it doesn't cause an immediate or direct change to the other particle's state. In the case of flipping the spin of one of the entangled particles, you are applying a unitary transformation to the entangled system. This will change the form of the entangled state, but it won't necessarily cause the spin of the other particle to flip. The effects of the transformation will only be apparent when both particles are measured. If you measure the spins of the particles after applying the transformation, you may find that the correlations between the measurements have changed as a result of the operation. This is because the unitary transformation you applied changed the nature of the entanglement. It's essential to remember that the act of measuring one of the entangled particles will collapse the joint quantum state and break the entanglement. After the measurement, any further operations on either particle will not affect the other. J H Can you give me an example of this experiment and the possible results? Gpt 4 Certainly! Let's consider a simple example involving two entangled electrons. Electrons have a property called spin, which can be either "up" or "down" along a given axis. We'll use the z-axis for simplicity. Prepare two electrons in an entangled state: Start by preparing the electrons in a maximally entangled state, called a Bell state. One such state is the singlet state, written as: |ψ⟩ = (1/√2)(|↑↓⟩ - |↓↑⟩) This notation means that the first electron (A) has spin up and the second electron (B) has spin down with equal probability as the first electron having spin down and the second electron having spin up. The (1/√2) factor is for normalization. Apply a spin flip operation to electron A: Now, without measuring the electrons, apply a spin flip operation (Pauli-X gate) to electron A. The Pauli-X gate acts on the spin state as follows: X|↑⟩ = |↓⟩ X|↓⟩ = |↑⟩ Applying the X gate to electron A will transform the entangled state as follows: |ψ'⟩ = (1/√2)(X|↑↓⟩ - X|↓↑⟩) = (1/√2)(|↓↓⟩ - |↑↑⟩) Measure the spins of both electrons: Now, measure the spins of both electrons A and B along the z-axis. There are four possible outcomes, with the following probabilities: A is down, B is down: Probability 1/2 A is down, B is up: Probability 0 A is up, B is down: Probability 0 A is up, B is up: Probability 1/2 So, when you measure the spins, you'll find that both electrons have the same spin with a probability of 1 (either both up or both down). Note that the correlations have changed compared to the initial state. In the initial state, if you measured the spins, you'd always find them to be opposite. This example illustrates how applying a transformation to one particle of an entangled pair can change the nature of the correlations between the particles without directly causing a change to the other particle's state. The effect of the transformation becomes apparent only when you measure the particles and observe the resulting correlations. J H Can you explain where a previous model got this explanation wrong?: "One example of an experiment involving entangled particles is the EPR (Einstein-Podolsky-Rosen) experiment. In this experiment, two particles are created in an entangled state, such that their properties are correlated with each other. For example, the particles might be entangled such that if one particle has a spin-up orientation along the x-axis, then the other particle will have a spin-down orientation along the x-axis, and vice versa. Now, suppose that we perform an operation on one of the particles, such as flipping its spin along the x-axis. This operation will cause the entangled state of the two particles to change. Specifically, after the operation is performed, the entangled state will be a superposition of two states: one in which the first particle has a spin-up orientation along the x-axis and the second particle has a spin-down orientation along the x-axis, and one in which the first particle has a spin-down orientation along the x-axis and the second particle has a spin-up orientation along the x-axis. The interesting thing about this experiment is that the state of the second particle is not fully determined until we make a measurement on it. If we measure the spin of the second particle along the x-axis, we will find that it is either spin-up or spin-down, with a probability of 50% each. However, if we measure the spin of the second particle along a different axis, such as the y-axis, we will find that it is either spin-up or spin-down, but the probabilities of each outcome will depend on the precise orientation of the measurement axis. This phenomenon is sometimes called "quantum nonlocality," because the state of one particle appears to be instantaneously affected by the operation performed on the other particle, even when the particles are separated by large distances. This is in contrast to classical physics, where such instantaneous correlations would be impossible. The EPR experiment has been the subject of much theoretical and experimental investigation, and it remains one of the most intriguing and mysterious phenomena in quantum mechanics." Gpt 4: The explanation provided in the previous response is not entirely accurate in describing the EPR experiment and the concept of quantum nonlocality. It does capture some essential ideas, but there are a few misconceptions and oversimplifications that need to be clarified. EPR experiment and entangled states: The EPR experiment was initially a thought experiment proposed by Einstein, Podolsky, and Rosen in 1935. Their goal was to question the completeness of quantum mechanics by pointing out its seemingly "spooky action at a distance." The experiment was later realized with actual particles, thanks to the development of Bell's theorem and experimental techniques. In the response, it is mentioned that the entangled state is a superposition of two states after flipping the spin of one particle. This is not incorrect, but the explanation lacks the clear distinction that the entangled state is a joint property of both particles and not a superposition of the individual states of each particle. Quantum nonlocality and measurement: Quantum nonlocality refers to the idea that entangled particles have correlations that cannot be explained by local hidden variables, meaning that the behavior of entangled particles cannot be described solely by the properties of each particle in isolation. This concept was further developed by John Bell, who derived inequalities (Bell's inequalities) that, when violated, demonstrate that the observed correlations are fundamentally quantum in nature. The previous explanation suggests that the state of the second particle is not fully determined until a measurement is made, which is misleading. In fact, the entangled state of the particles is a joint property, and once a measurement is made on one particle, the entangled state collapses, and the correlations between the particles become apparent. The measurement on the first particle does not "instantaneously affect" the state of the second particle. Instead, it merely reveals the correlations that were already present in the entangled state. Probability of measurement outcomes: In the given example, it is mentioned that the probabilities of the measurement outcomes along the x-axis are 50% each, while the probabilities along the y-axis depend on the precise orientation of the measurement axis. This is somewhat misleading because it seems to imply that the probabilities change depending on the measurement basis. The probability of getting a specific outcome depends on the entangled state and the chosen measurement axis. If the entangled state and the measurement basis are changed, the probabilities of the outcomes may also change. Overall, the previous response captures some general ideas about entanglement and quantum nonlocality but lacks precision and clarity in explaining the concepts and their implications accurately. Much more detailed response, and it could even tell where the last model got things wrong!

  • @umbrakinesis2011

    @umbrakinesis2011

    Жыл бұрын

    It also gets spacial relationships right (ie: is this place farther north than this one), and seems to understand 4d geometry decently well.

  • @miroslavcermak9835
    @miroslavcermak9835 Жыл бұрын

    Hello Sabine and the internet. 19:12 "old fashioned thing called reality" :D Hilarious, just hilarious, not unlike other parts of the video. But the video is also very informative and interesting, like the explanation of understanding as a creation of a model . So thanks for the work put in!

  • @soasertsus
    @soasertsus Жыл бұрын

    This topic is so interesting to me and I agree with you, I think it's very clear that GPT has some sort of rudimentary understanding. It's just an extreme savant, it understands language very well and using that ability it can cobble together models of some other basic areas it wasn't built to understand like coding or basic math, like how a human with a traumatic brain injury that damages some vital brain function can sort of press gang other parts of their brain to build new methods of accomplishing something similar or how a person who is blinded might be able to be able to rewire their other senses to compensate to a degree. But it fundamentally just lacks the neural circuitry and information to understand more complex things about the real world and lacks the "sensory organs" if you will to make observations about the outside world and check its models against them because it's doesn't have the ability to search the internet or interact in the physical space, which is why for some topics all it can do is make a best guess using the limited information it has and often gets it wrong. Again it's like asking a blind person what things look like: they can guess based on what they've heard described and logic but they fundamentally just don't have the ability to know and even if they guess right they're not capable of actually understanding what it's even like to see. But ChatGPT almost never make a mistake at what it's specced for, which is generating grammatically correct and believable text. Additionally, I think a lot of humans are on an unjustifiable high horse when it comes to consciousness. People want to believe we're special when any sober and rational analysis would show that we are most likely operating just like these AI systems, but with more complexity, and that consciousness is just an emergent property of that complexity, or an abstracted model of the computing hardware itself inside that hardware which must be useful for some purpose. I'd very much suspect that if we discovered the mechanism, it would be possible to turn off the consciousness in a human brain with some targeted intervention and turn us back into unaware computing machines. Hell, similar states are possible with drugs for example. We shouldn't be so quick to dismiss AI, or other animals even, because the fact is they operate on the same principles as us, the difference is just our hardware is much more advanced and specialized to different tasks compared to silicon computers.

  • @nnnik3595

    @nnnik3595

    Жыл бұрын

    What will be interesting is GPT-4 then which promises to be basically gpt-3 but with multiple sensory mechanisms (It can read text, see images and hear sounds)

  • @C0Y0TE5

    @C0Y0TE5

    Жыл бұрын

    The purpose of, and research into anesthetics is to turn off conciousness.

  • @katrinabryce

    @katrinabryce

    Жыл бұрын

    I tried this experiment: I asked Bing Chat to give me directions from Real London to Fake London. It said it didn't understand what I meant by "Fake London", so I told it I meant London Ontario. It then told me they are 5500km apart, and to visit Google 🤯 Maps to get directions 🤷🏻‍♀. Then I asked it if Fake London is to the north of Real London. It once again didn't understand "Fake London" even though I had just told it what I meant. Then when reminded, it made the same error about it being further north than the real one.

  • @madattaktube

    @madattaktube

    Жыл бұрын

    Agreed, I have often viewed conciousness as 'If it walks like a duck and it talks like a duck...' - if by all observations it acts just like a concious being, then it *is* a concious being, be it a Chinese Room or not. If there is no meaningful way to distinguish them, how can they really be two different concepts? Obviously I think ChatGPT is still a long way off. While it has blown all previous AI models out of the water, auto-complete on steroids still seems to be the best descriptor of its behaviour IMO.

  • @AORD72

    @AORD72

    Жыл бұрын

    Well said. I think it shows us that we a close to machines that will be superior to us and that will supersede us.

  • @violoncello4439
    @violoncello4439 Жыл бұрын

    7:16 This also misses the point of the Chinese Room. The room is indeed an excellent model of translation, so if you define understanding as creating models, then the room certainly has understanding. However, the original question is whether _any element_ of the room knows the _meaning_ of Chinese words. The answer is no; a rule book has been written for the manipulation of _syntax_ but _semantics_ was not imparted to the room. Chinese speakers know what Chinese words refer to, while the Chinese Room only arranges Chinese words in a way _they interpret_ as correct. The Room has absolutely no knowledge of Chinese. It is a mechanism for manipulating symbols in a way that is meaningful to _us_ , but no element of the system knows the meaning of Chinese words. Hossenfelder’s definition does not account for meaning, which is in fact the central point of contention in the Chinese Room, not just ‘the creation of models’!

  • @PlasticPaddyFalun
    @PlasticPaddyFalun Жыл бұрын

    Thank you for a great channel! I rarely leave comments anywhere but I feel that I must give a nuance to your discussion by pointing out that in Searles Chinese room thought experiment both the input and the output are in Chinese (he choose that language because he had no knowledge of it whatsoever). Searle - or the person in the room - represents the CPU and the book of instructions the represents the program running in the computer. The person in the room understands the instructions but neither the input or the output. Searles point is that computers understand syntax but not semantics; the process but not the meaning. Searle also notes that the Universal Turning Machine (UTM) is only capable of a very limited number of operations, write, move, state transition, branching, halting. All digital computers are in essence instances of this UTM. In terms of large language models the program is trained to break down words (lemmatization) into tokens which contain numerical vectors that represent the meaning of words. The AI performs mathematical operations using these vectors, takes an input and returns an output. The program, just as the person in the Chinese Room is only ever performing the mathematical operations (syntax) it has been programmed/trained to do. The machine is incapable of that much higher level activity we call understanding (semantics). I am of the opinion that we will not get the most out of the fantastic tools that AI can offer us if we make a fundamental mistake about what AI can and can't do. John Searles argument is as powerful today as it was when he first presented it to the world in the early 80s. That's what philosopher should be good at: thinking about how the world really is! Thank you once again for you excellent channel.

  • @dirkk9931
    @dirkk9931 Жыл бұрын

    9:55 gave me a bit goose bumps- scary 😱 Thanks for your video. Do you think if AI understands that it might also have consciousness? Maybe we are also only biological robots but every complex structure has some kind of consciousness?

  • @fandomguy8025

    @fandomguy8025

    Жыл бұрын

    My favorite theory of consciousness is the attention schema or attention model theory. It posits that consciousness is a result of our brain creating a model brain in order for it to understand and thus control itself. The brain needs to focus on some information but not all the information it gets, generates, or stores because it's just too much. As you read this, you ignore everything else going on around you and inside you. This is it's attention system, and the attention model helps the brain control it. Since models are useful but simplified descriptions, it takes the form of a "spirit" or "agent" that's in the machine, instead of the ultracomplicated mess of neurons the brain actually is. This is what we (we being our brains) call the soul. That magic thing inside the meat robot that can focus on stuff (Sights, sounds, smells, memories, feelings, etc.) (These models are also simplified. "Color" isn't really a thing, light has different wavelengths, it's just a simplified model your brain uses to show that not all light is the same.) Though, of course, it's more complicated than that since our attention model also connects to our brain's body model and that connects to our brain's world model so we perceive ourselves as magic things that focus on stuff + located in the head of a body that senses stuff + are somewhere in the world. (And of course, our brain's learned language model, allowing you to talk like a chatbot.) It'd be very scary to just be an attention model existing in zero gravity darkness and quiet, what sensory deprivation tanks try to achieve. (Causing vivid hallucinations since the brain does not like having no input and will automatically generate stuff. People consider this fun.)

  • @MoominPa
    @MoominPa Жыл бұрын

    I think what we always assume that “understanding” involves some degree of consciousness. We are in a new territory. We might need a new word for “Understanding” when it comes to AI.

  • @mikesawyer1336

    @mikesawyer1336

    Жыл бұрын

    I think back to my early childhood, where I understood some things, but as I recollect and compare to my conscious level of interacting in the world today, I would argue that in the beginning I was not conscious. Not really. At what age did I attend to my independent needs? I think about this a lot. At some point I woke up and became conscious.

  • @marcomoreno6748

    @marcomoreno6748

    Жыл бұрын

    Will "consciousness" become the new "soul"? I wouldn't write it off -Marco

  • @greenanubis

    @greenanubis

    Жыл бұрын

    Im starting to think that consciousness is just another word for agency. We apply it to the things that do stuff without being told to do it. I think that the whole thing just arises from our biological instincts, consciousness itself and our recognition of it. Part of it is defining consciousness as something special that other humans have so we can function in human society. All that doesnt really have anything to do with does that agent really understand anything or not, or with understanding itself.

  • @rickfrombohemia9550

    @rickfrombohemia9550

    Жыл бұрын

    Yeah, it's not about understandinng as such, but about reflection of that understanding, you know/think you understand something (whether right or wrong is irrelevant). You even don't have to understand anything in order to be a conscious being. And vice versa, AI doesn't have to be conscious in order to understand something.

  • @thomassturm9024

    @thomassturm9024

    Жыл бұрын

    @@greenanubis Gold. Every single word of it. This is by far the best definition of "consciousness" I have ever come across.

  • @shtthemnky
    @shtthemnky Жыл бұрын

    Love you video Sabine. Given that vision grabs 10x more information than listening... I'd expect multi-modal LM to explode in their abilites. I wait for your vision/interpretation of emergence in this and future models!

  • @herbertdarick7693

    @herbertdarick7693

    Жыл бұрын

    Well, I prefer language over graphs and other visual material, so it depends on who is your target audience.

  • @omnicatalyst
    @omnicatalyst Жыл бұрын

    I'm working on a D&D campaign and built a "boon" system with five levels of perks. I wrote up 3 samples, fed them to chatgpt, and it was able to reliably put out creative iterations that were mostly balanced

  • @saravananrajendran4170
    @saravananrajendran4170 Жыл бұрын

    You may be very surprised Sabine when you put the same prompts re: entangled particles into GPT-4 instead. It gave, what would be the correct answer here (pasted below!). This makes me think the observations you have shared in this video, aren't a fundamental feature of transformer models, but just a limitation of GPT-3 which was overcome. Q: If you perform an operation which is not a measurement on one particle (like flipping spin), in a pair on entangled particles, does that affect the other particle? When you have a pair of entangled particles, their properties become correlated in such a way that the state of one particle is dependent on the state of the other, even if they are separated by large distances. This correlation is described by the entanglement itself, and it persists until the entanglement is broken, usually due to a measurement or interaction with the environment. A: If you perform an operation on one of the entangled particles, such as flipping its spin, you are indeed modifying the state of that particle. However, the effect on the other particle is not as direct as it might seem. The entanglement between the particles means that the state of the second particle is still correlated with the first, but it doesn't mean that the state of the second particle changes instantaneously or deterministically as a result of the operation on the first particle. To illustrate this, let's consider a simple example of a pair of entangled particles with spins. Suppose that their spins are entangled such that when one is measured to be "up," the other will always be "down," and vice versa. If you flip the spin of the first particle, you have changed its state, but you have not directly changed the state of the second particle. The entanglement between the particles still dictates that if you measure the spin of the first particle and find it to be "up," you can be certain that the spin of the second particle is "down." In summary, performing an operation on one particle of an entangled pair, like flipping its spin, does not directly or instantaneously affect the state of the other particle. Instead, it modifies the relationship between the particles, which remains correlated due to their entanglement.

  • @IngoSchwarze

    @IngoSchwarze

    Жыл бұрын

    Wait, that still isn't the correct answer unless i got my elementary quantum mechanics very wrong. Initially, the state of the system is |1+>|2-> + |1->|2+> (modulo normalization and non-physical phase factors). If you now flip the spin of the first particle, the state changes to |1->|2-> + |1+>|2+>. If you now measure the state of the first particle and find it to be |1+>, you then know that the state of the second particle is |2+>, too, because |2-> + |1+>|2+>) = |2-> + |2+> = 0|2-> + 1|2+> = |2+>. To summarize: Wrong statement by ChatGPT: if you measure the spin of the first particle and find it to be "up," you can be certain that the spin of the second particle is "down." This is correct instead: if you measure the spin of the first particle and find it to be "up," you can be certain that the spin of the second particle is also "up." This means ChatGPT still has a wrong model and still does not understand (this kind of) elementary quantum mechanics. By the way, this illustrates perfectly what Sabine said: Theoretical physics is often easier to explain in mathematical formulae than in words. That does not just apply to quantum physics, by the way, you can see the same effect in classical theoretical mechanics, classical theoretical electrodynamics, and classical relativity theory.

  • @brothermine2292
    @brothermine2292 Жыл бұрын

    If you define "understands" as "can use with few errors" then it's fair to say that chatbots understand language. But that's a low bar. For one thing, it doesn't imply an understanding of language semantics. And it of course doesn't imply subjective experiencing of the meaning of its own sentences.

  • @SabineHossenfelder

    @SabineHossenfelder

    Жыл бұрын

    The relevant point is that you have a model.

  • @jhoughjr1

    @jhoughjr1

    Жыл бұрын

    @@SabineHossenfelder The relevant point is it gets an A on the test yet doesnt know how to apply it. You are really stretching the definition of understanding to suit your conclusions.

  • @brothermine2292

    @brothermine2292

    Жыл бұрын

    @@SabineHossenfelder : The relevance of having a model depends on the definition of "understands." Also, in the video a "model" is just a poorly defined black box.

  • @jonbbbb

    @jonbbbb

    Жыл бұрын

    @@jhoughjr1 what do you mean that it doesn't know how to apply it? Chat GPT is proving it applies the rules it learned during training every day in interactions with real people. Honestly confused about what you're trying to say.

  • @jonbbbb

    @jonbbbb

    Жыл бұрын

    @Brother Mine saying that using language with few errors is a low bar is really hard to believe. It's an incredibly high bar. One might say it's the highest bar we know how to set that real people can also achieve. It's the basis of the Turing Test for intelligence for example. You've made me curious to hear what you think would be a higher bar to demonstrate understanding than the full generality of human language. But even in typing that out, I've realized you're describing a process that by definition you can't demonstrate to me because that would use language.

  • @5irefly
    @5irefly Жыл бұрын

    In Searle’s Chinese Room, the rule book is the algorithm+data, not necessary just a lookup table. He’s going after the Turing Test which argues that an AI that produces human-equivalent input-output must actually understand the input. However, if the AI is merely performing a mechanical symbol substitution, no matter how sophisticated, then it arguably can produce satisfactory outputs without necessarily understanding the input. I read his book back in the early 80’s and that’s what stuck with me. (I went on to study ANN’s in grad school and use them commercially today, so do understand various algorithms , training, back prop, etc.)

  • @ralphmacchiato3761

    @ralphmacchiato3761

    Жыл бұрын

    Watch the video first, please.

  • @nathanharrenstein1298

    @nathanharrenstein1298

    Жыл бұрын

    I would like to thank you, Parseeker, for bringing this distinction up and I would like to expand on it slightly. Another way to put Parseeker's point is that Searle is pointing out the distinctions between syntax (the rules based structure of a language) and semantics (the meanings of the words). Searle's main assertion with the chinese room is that syntax does not get us to semantics. Parseeker rightly points out that look-up tables are only one form of algorithm that Searle is after, his stated target was "hard AI" or the idea that digital systems (or any other symbol manipulation methods) establish the necessary conditions for consciousness. Sabine is a very intelligent presenter and a good communicator, but held within her presentation are a couple of assumptions that It would be irresponsible not to point out. Sabine's point turns on the definition of the term model and whether or not we consider the model as having any sematic content. She assumes model can process semantic content or possibly the syntactic structure can create semantic content. This assumption seems to come out of a metaphysics of (reductive?) materialism which seems to be her other assumption. I have sympathies for these positions and believe they are perfectly reasonable even if I do not share them. When discussing AI Models I am less convinced than Sabine that the training contains semantic content and even more skeptical of the idea that the rules of the algorithm could create it. Frankly put, If there is no meaning to start with, I do not see how rules about the probable the next best word can give us the meaning of the current word. Like Sabine points out we are not sure if this is anywhere close to how we learn because we only can assess the inputs and outputs. Although I do not share her possition, or Searle's for that matter, I appreciate Sabine engaging in this conversation and enjoy her content.

  • @kakistocracyusa

    @kakistocracyusa

    Жыл бұрын

    @queerdo "she's presenting her metaphysical position as a scientific one. That is dishonest." This nails it - is 100% correct and mirrors my same comments on her videos. I've come to conclusion that she's is incentivized to remain so intellectually dishonest.

  • @ArawnOfAnnwn

    @ArawnOfAnnwn

    Жыл бұрын

    ​@@kakistocracyusa It's not dishonest to present a topic as you understand it. She sees the world through a materialist lens and explains it as such - that is being true to her viewers about what she knows. And you can't say how she views these things is wrong cos the issue is still open in both science and philosophy.

  • @kakistocracyusa

    @kakistocracyusa

    Жыл бұрын

    @@ArawnOfAnnwn You assume she understands it, which she doesn't - so then, like you, she is simply ignorant of when she is discussing metaphysical questions and when she is discussing actual physics; and obfuscates her own ignorance of both physics and classical philosophy by deflecting to irrelevant quantum mechanical mysticism: the standard escape route for sub-standard physicists who opt for being pop-stars instead of productive physicists (see Fritjof Capra "The Tao of Physics" for another example of such reliance on spineless equivocation through muddy mysticism).

  • @AccipiterPictures
    @AccipiterPictures Жыл бұрын

    I appreciate all that this video unpacks! I think something worth distinguishing is an entity having the ability to model meaning and an entity having the ability to self-consciously experience modeling meaning. I'm not sure there's evidence yet to suggest that these chatbots can self-consciously experience anything even if their structures and ability to self-update permits them some form of understanding or even evolving understanding. Curious of course if anyone disagrees.

  • @jackhill2765
    @jackhill2765 Жыл бұрын

    Great video, particularly pointing out the differences between the interior architecture of chat GPT and the Chinese room. I watched this several times before realizing what it was that for me was missing from these discussions of whether AI can understand. Namely, in sentient humans, understanding has a subjective component i.e. the "aha moment" of "I understand", that is the subjective feeling (qualia) that accompanies understanding. Understanding could be defined as requiring this subjective element in addition to the objective (scientific?) observable parts including the ones related to the interior architecture of the bot, or not i.e. you could define understanding as not requiring the subjective quality. In any case for me, similar to experiencing the color red, the subjective "aha moment" is the essence of understanding.

  • @IngoSchwarze

    @IngoSchwarze

    Жыл бұрын

    Most definitely it is not. While i'm not a professional teacher or professor, i have worked in education in very different contexts (chess at schools, physics and mathematics at universities, and training youths at mountaineering clubs). It happens regularly that people suddenly cry out "oh wow! now i finally understand" and then when you go and check, they very clearly still do *not* understand the aspect they just had their heureka moment about. The reverse does happen, too: that somebody feels very sceptical about their own understanding (in the words of Sabine: as if something is missing in their own model in their own mind), and yet, they are able to apply their internal model extremely well to all kinds of unknown problems, sometimes even better than all the rest of the group, and yet they feel as if they still lack their heureka moment. According to my experience, teaching humility to the boastful, overconfident and excitable and teaching self-confidence to the timid and sceptic takes up a significant part of time and in particular of emotional effort in teaching. Also, judging the level of your own understanding of a topic is surprisingly hard and needs a lot of training and practice.

  • @mrubengmail
    @mrubengmail Жыл бұрын

    So much has been written about chatbots recently, but this is the best video (or popular explanation of any kind) that I’ve seen. Really appreciate how you provide a clear, non-specialist set of explanations and analysis without sacrificing conceptual or theoretical precision. Nicely done, and thank you! P.S. as I type this comment on my phone, I find myself eagerly looking forward to the day chatbot AI is integrated into autocorrect algorithms. :-)

  • @aguyontheinternet8436

    @aguyontheinternet8436

    Жыл бұрын

    Or replace it with the auto correct that search engines use. Search engines are good at correcting words, just copy from them!

  • @3DisFuntastic
    @3DisFuntastic Жыл бұрын

    As always, extremely interesting and fun to watch.

  • @codebycarlos
    @codebycarlos Жыл бұрын

    Already ChatGPT is improving. I asked it with version 4 the same exact latitude question. This time I got: Windsor, UK is further north than Toronto, Canada. Windsor is located at approximately 51.48 degrees north latitude, while Toronto is located at approximately 43.70 degrees north latitude.

  • @S13Reborn
    @S13Reborn Жыл бұрын

    14:25 This is what you get from that same prompt with GPT-4: The entanglement of two particles, as described by quantum mechanics, implies a deep and fascinating connection between them. When two particles are entangled, the state of one particle is immediately connected to the state of the other, regardless of the distance between them. This doesn't mean that information is transmitted faster than light, but rather that the measurement outcomes on both particles are correlated in a way that can't be explained by classical physics. However, this correlation only appears when a measurement is made. If you perform an operation on one of the particles that does not involve measurement, the other particle will not be immediately affected or altered in any way. The quantum state of the system might change due to the operation, but the entanglement won't manifest itself until a measurement is made. Moreover, if the operation performed is unitary (a standard type of quantum operation that doesn't involve measurement), it will generally just change the form of the entanglement without breaking it. For instance, if the particles were entangled in the position basis, a unitary operation might change this to entanglement in the momentum basis. If the operation is not unitary (e.g., involves some sort of decoherence), it could potentially destroy the entanglement. It's important to remember that entanglement doesn't mean that one particle is sending signals or influencing the other instantaneously. Rather, entanglement creates correlations between the measurement outcomes, which become apparent only after the measurements are compared.

  • @o_o8203
    @o_o8203 Жыл бұрын

    When humans learn how to speak, the adults around us identify things for us, ask us questions about things that they've identified, correct our answers when they're incorrect, and give us positive feedback when our answers are correct. This seems not so different from how chatGPT learns. This is why I always say thank you to the chat bots 😂

  • @polarizedpotstone

    @polarizedpotstone

    Жыл бұрын

    I think I‘d disagree. GPT3, the actual model behind ChatGPT, has been trained on a whole ton of internet. Humans learn piece by piece in their own limited bubble of life. Also, reinforcement learning (equivalent to our surroundings correcting us) is only one technique to train an AI. GPT3 (as far as I understand) has not been trained with reinforcement learning, but really just got a big af info dump. OpenMind however does use reinforcement learning to further align GPT3, resulting in models like ChatGPT, so I do have to somewhat agree there. Oh and I also can‘t help to alway say thank you :D

  • @notanemoprog

    @notanemoprog

    Жыл бұрын

    I will remember to say "thank you" when the ChatGPT robots come to take me away to a processing plant

  • @amihart9269

    @amihart9269

    Жыл бұрын

    You can train a neural network through a process like you described, where a "teacher" tells the neural network if its decisions are correct or incorrect. But you can also train a neural network through evolutionary algorithms, by creating replicas of the network with slight variations and getting rid of the ones that perform worse and keeping the ones that perform better. People tend to think of humans as only learning from birth until death, but in life as a whole has been learning through biological evolution for hundreds of millions of years. You are born with a lot of pre-trained brain structures programmed into your DNA so that your body builds these brain structures as you grow up. For an AI, it's much more difficult to learn even simple things like how to recognize objects in images because it does not have the benefit of hundreds of millions of years of pre-training evolution, while your visual cortex is partly something trained through reinforcement learning and partly something constructed from DNA information you inherited.

  • @dtibor5903

    @dtibor5903

    Жыл бұрын

    ​@@polarizedpotstone it's trained on dataset scraped from internet and other sources but they were selected, curated and labeled by hundreds of humans...

  • @bartsanders1553

    @bartsanders1553

    Жыл бұрын

    It doesn't have physical extrenal stimuli, though.

  • @Embassy_of_Jupiter
    @Embassy_of_Jupiter Жыл бұрын

    It's interesting watching these models write something. For every word it writes, you can look at what the attention mechanism is doing. It more or less shows "what words were important to generate this next word of output". It does seem very human like in how it references previously written words or words in the prompt. You can pretty much see its train of thought.

  • @Embassy_of_Jupiter

    @Embassy_of_Jupiter

    Жыл бұрын

    @@aliceberethart Yeah from what I've seen this seems to be a good approach to AI, because it can never think without being prompted, it can never learn unless specifically being trained and it seems to be pretty "empathetic" in that it can replicate speech from the point of view of anyone, in any state of mind, including emotions.(that's not as obvious in ChatGLP, but it is in GPT-3, because it wasn't specifically trained to play the role of the helpful emotionless assistant) Doesn't mean that it's feeling empathy of course, but it doesn't mean that it has the capacity to understand us perfectly. But by default you can tell it to be the biggest psycho possible or tell it to be a #empath. You could very easily punish non-empathic behavior and reward empathic behavior during pre-training/fine-tuning/zero-shot-learning. In this way you could create a model that acts like the most empathic person on the planet, just like they trained ChatGPT to be neutral and helpful. But of course you could just as easily create angry-13-year-old-on-Xbox-live-GPT. In a way we are lucky that this is the first way we found to make usable AI, because it seems inherently rather safe an controllable. I could envision other architectures where that's not the case and it just learns by exposure uncontrollably with a stream of consciousness.

  • @polarizedpotstone

    @polarizedpotstone

    Жыл бұрын

    @@aliceberethart I think that‘s a pretty dangerous misunderstanding: ChatGPT has not somehow magically evolved to 'think' the same way we do, rather it has been trained on data that has been created by humans that think like humans so there is no other way for it to 'evolve' other than similar to our own thinking patterns.

  • @minimal3734

    @minimal3734

    Жыл бұрын

    @@aliceberethart I agree. And I think it's fair to say that the AI is essentially human in that sense.

  • @dawn21stcentury
    @dawn21stcentury Жыл бұрын

    What a brilliant video; the insights, the dry humor, absolutely brilliant 😂❤🎉 So I also just 'changed' my model of the world: Since I am Dutch, I always thought Germans didn't have humor. I now stand corrected 😁

  • @joshlanders
    @joshlanders Жыл бұрын

    Very good discussion and viewpoint!