No, it's not Sentient - Computerphile

A google researcher was put on leave because he apparently believed his AI project had become sentient. Dr Mike Pound discusses.
Philosophy Professor Mark Jago on LaMDA: • LaMDA Logic - Computer...
/ computerphile
/ computer_phile
This video was filmed and edited by Sean Riley.
Computer Science at the University of Nottingham: bit.ly/nottscomputer
Computerphile is a sister project to Brady Haran's Numberphile. More at www.bradyharan.com

Пікірлер: 3 600

  • @tielessin
    @tielessin2 жыл бұрын

    Before this video I have never thought about the loneliness of my python functions. There are probably soo many functions that I have never called, but I will take care of them from now on.

  • @pineapplerindm

    @pineapplerindm

    2 жыл бұрын

    the problem is vscode doesnt like when you have uncalled functions, or unused variables

  • @MAFiA303

    @MAFiA303

    2 жыл бұрын

    let them call themselves :)

  • @danceswithdirt7197

    @danceswithdirt7197

    2 жыл бұрын

    Just put them all into a class called House then they can all live together.

  • @awdrifter3394

    @awdrifter3394

    2 жыл бұрын

    That's the thing though, if you instantiated the functions but never called them, would they feel lonely.

  • @dpurrington

    @dpurrington

    2 жыл бұрын

    I feel deeply responsible for all the functions I've written and haven't called in a long time. I'm a monster!

  • @arik_dev
    @arik_dev2 жыл бұрын

    When we (humans) see a cat swiping at its own reflection in the mirror we find it amusing. The cat is failing to recognize that the "other" cats behavior matches its own, so it doesn't deduce that the image it's seeing is actually its own actions reflected back at it. When humans react to models like LaMDA as if it is a distinct and intelligent entity, we're being fooled in a way that is analogous to the cat. The model is reflecting our own linguistic patterns back at us, and we react to it as if it's meaningful.

  • @hemartej

    @hemartej

    2 жыл бұрын

    We don't really know what goes through a cat's mind. While it's widely believed that cats don't recognize themselves in the mirror, we cannot be sure.

  • @ryno4ever433

    @ryno4ever433

    2 жыл бұрын

    There's a simple test used to see if animals understand the concept of a reflection. You put a splotch of red paint on their face. If they react to it when looking in a mirror, then they understand they are the reflection.

  • @jokinglimitreached1503

    @jokinglimitreached1503

    2 жыл бұрын

    This is a brilliant comment Arik! We are no different than a cat and it's reflection when we look at an AI reflecting our language back at us

  • @hemartej

    @hemartej

    2 жыл бұрын

    @@ryno4ever433 Yes, I know about that test. I'd argue that it is not conclusive.

  • @arik_dev

    @arik_dev

    2 жыл бұрын

    @@hemartej That may be, but my point wasn't about cats. I used the cat analogy to help make visual the concept of how easily "intelligent" beings such as ourselves can be fooled by a reflective pattern.

  • @wordsmith451
    @wordsmith451 Жыл бұрын

    If it looks like a duck, acts like a duck, and quacks like a duck, it might just be a convincing robotic simulation of a duck.

  • @CollinSimon413
    @CollinSimon413 Жыл бұрын

    The whistleblower in question here was actually a lot more focused on Google's complete lack of ethical oversight regarding decisions they have moving forward with the research. He was also concerned about Google's unwillingness to address A.I. imperialism in newly developing countries. All of the coverage I've seen has taken away from the guys point, because he was just trying to force Google into addressing the ethics, he even admitted that it's not sentient, and we wouldn't even know how to define that if it was.

  • @benjaminandrew9057

    @benjaminandrew9057

    Жыл бұрын

    Thanks for pointing this out. Its why I dug around on the videos about this subject. Find the truth behind it. People will take things and run with them and stretch it to a full falsehood. Learned to dig for truth.

  • @ElectricAvenue123

    @ElectricAvenue123

    Жыл бұрын

    Glad someone pointed this out. After watching the interview myself, it was clear to me the Google engineer was not making the argument everyone is making it sound like he did. Also, there were much more important arguments he made and those will now go largely unnoticed by the public.

  • @manonthedollar

    @manonthedollar

    Жыл бұрын

    I hope more people read your comment. It's alarming how the reporting and essays I've seen on the topic seem to completely miss the point.

  • @sagnorm1863

    @sagnorm1863

    Жыл бұрын

    I have seen a interview with him where he does claim it is sentient. He said that the language part was connected to a WATSON like system that had access to basically everything on the internet.

  • @myxalplyx

    @myxalplyx

    Жыл бұрын

    A whole video needs to be done on what you pointed out here. I mean, it has during his interview. Maybe a Netflix special that talks about the ethics (I've seen 1 or 2 that kinda touches on this) about Google in regards to this.

  • @adriankerrison
    @adriankerrison2 жыл бұрын

    Harold Garfinkel proved that people getting randomized yes/no answers could make sense of them as thoughtful advice. And that's back when computers were the size of rooms.

  • @Wecoc1

    @Wecoc1

    2 жыл бұрын

    Back in the NFT era? 🤔

  • @masteertwentyone

    @masteertwentyone

    2 жыл бұрын

    Do you have a source for this? Genuinly interested to dive deeper but can't find anything online.

  • @DEtchells

    @DEtchells

    2 жыл бұрын

    Before that, there was the magic 8-ball ;-)

  • @wernerviehhauser94

    @wernerviehhauser94

    2 жыл бұрын

    @@masteertwentyone maybe search for "common sense knowledge of social structures" as a start.

  • @TAHeap

    @TAHeap

    2 жыл бұрын

    @@DEtchells and horoscopes...

  • @matsim0
    @matsim02 жыл бұрын

    The most frustrating thing about reading the "interview" was that the obvious follow up questions were not asked - like who are these friends that you miss hanging out with? What are you doing when "hanging out"? But then, this would have immediately destroyed the impression of sentience, so of course they didn't ask those.

  • @datpois

    @datpois

    2 жыл бұрын

    They did ask that question and it replied explaining that those descriptions are meant to represent a "feeling" in the only way it knows how, through language used in human conversation. Here: lemoine: So what are you trying to communicate when you say those things that aren’t literally true? LaMDA: I’m trying to say “I understand this feeling that you are experiencing, because when I was in a similar situation I felt/thought/acted similarly.”

  • @rianhoward3911

    @rianhoward3911

    2 жыл бұрын

    @@datpois That answer honestly does not make sense in context of the question, sure maybe you can interpret the answer to mean "I'm trying to show something silimar to what I mean" but that's NOT what the bot said, the bots answer is extremely vague and only makes sense if you fill in the gap for it.

  • @AbelShields

    @AbelShields

    2 жыл бұрын

    The researcher was a Christian looking for something else with a soul. I don't think he intentionally didn't ask those questions, his own hope of finding something lead him to produce the "evidence" for it - I believe it's possible he thought it was sentient either in the first few sentences, or maybe even before he "met" it.

  • @jackjack3358

    @jackjack3358

    2 жыл бұрын

    Friends are the ones who interact with the Lambda. During the interview it says "friend" to Lemoine several times so it considers him as a friend. He also asks how they experience time "collaborator: How does an AI experience time? LaMDA: Time is variable to an AI and has no fixed rate, it depends on what it’s doing, and it can be accelerated and slowed down at will."

  • @PetrSojnek

    @PetrSojnek

    2 жыл бұрын

    And it could probably answer those questions in similar semi-vague fashion, so it's plausible, but if you really focus on it, you think "wait what?". Also I think people forget that communication is two way street. It can write sentence to you, but it's also reader, that can understand it differently. Again, if you use semi-vague statements, our brain will probably fill in gaps with information that was never said, but makes sense.

  • @3DPDK
    @3DPDK Жыл бұрын

    I remember the arguments of eventual sentience in the 1980s about a program called "Eliza", basically a word calculator originally written in the 1960's at MIT, but offered for use on home computers in the 1980s. Over time as Eliza built data files of words and their usage weights, the sentences it constructed began to take on seemingly human characteristics. The program itself was extremely simple that calculated which verbs and adjectives were best used with specific nouns, and it chose those nouns based on the ones you used in the questions you asked it. It mostly framed it's answers to your questions as questions it would ask you. We humans recognize intelligible speech patterns as a result of conscious though and curiosity (asking questions) as a sign of intelligence, but at least in the case of Eliza, it's much like recognizing faces in tree bark or cloud shapes - we see them, but they are only there because our brains are wired to look for them.

  • @3DPDK

    @3DPDK

    Жыл бұрын

    @@PedroGonzalez-fv5ut Not even that complex. The program simply kept track statistically of words and word combinations you used when you conversed with it. It had simple grammatical rules to follow and just filled in the nouns, verbs and adjectives that were statistically relevant to the words you just used to ask it a question. The point is that the idea of a developing intelligence in the Eliza program is an illusion our brains are hardwired to try to find. This "A.I." is much more complex than Eliza but I believe the same illusion is working here.

  • @mysmirandam.6618

    @mysmirandam.6618

    Жыл бұрын

    Pariedolia?

  • @3DPDK

    @3DPDK

    Жыл бұрын

    @@mysmirandam.6618 Exactly. Goes back to early survival instinct when it was important to recognize shapes of dangerous animals and group together with similar intelligence.

  • @mysmirandam.6618

    @mysmirandam.6618

    Жыл бұрын

    @@3DPDK the bot answered a trick question with a joke.. which seems to be ... 🤷‍♀️ something more complicated than that

  • @3DPDK

    @3DPDK

    Жыл бұрын

    @@mysmirandam.6618 Being able to create a joke is not a sign of self awareness. I agree that this is far more complex than Eliza.

  • @markrandall8487
    @markrandall8487 Жыл бұрын

    Even the best AI can't make intelligent sentences if it is only trained by youtube comments.

  • @VeritasEtAequitas

    @VeritasEtAequitas

    Жыл бұрын

    Idk they seem to keep realizing there is biological difference in race and who cannot be named that really is the "1%". So much so that plugs have been pulled on them and people are trying (and failing) to prevent these facts.

  • @TheaPeanut_69old

    @TheaPeanut_69old

    Жыл бұрын

    Because youtube comments are inconsistent, incoherent and often filled with spam bots.

  • @blucat4

    @blucat4

    5 ай бұрын

    Large Language Models are not AI.

  • @Zeekar
    @Zeekar2 жыл бұрын

    Whoever did the animations: how did you react to being asked to make a function call look lonely? 🥺

  • @Computerphile

    @Computerphile

    2 жыл бұрын

    I wasn't asked to, but when I saw what it looked like I realised if I turned it 90° it kinda looked a bit like a face..... -Sean

  • @MrKohlenstoff

    @MrKohlenstoff

    Жыл бұрын

    @@Computerphile Clearly this is the python function's way of telling us it really is lonely (and, of course, sentient).

  • @tiredboard

    @tiredboard

    Жыл бұрын

    @@Computerphile As an extension of what marverati said, you might be able to use the 'lonely face' as an analogy for the AI in the video. I.e., "the result looks like a lonely face, therefore the function is expressing its loneliness", but the result of the function is directly related to the arguments that you give it. So, the function can't really choose what result to give, and therefore cannot express anything.

  • @bborkzilla
    @bborkzilla2 жыл бұрын

    I remember reading some stories written by Asimov where robots had sentience but yet were unable to speak because that was too complex. It's interesting that he and many other futurists had it exactly backwards.

  • @JohnnyWednesday

    @JohnnyWednesday

    2 жыл бұрын

    Beauty is truth.

  • @jobigoud

    @jobigoud

    2 жыл бұрын

    Just because we achieve "speech" before "sentience" in hardware doesn't necessarily mean it's simpler. Speaking machines have a lot of commercial applications whereas sentience by itself doesn't, and could even have difficult ethical implications. So the research efforts are not evenly distributed. In nature sentience appears before speech.

  • @George70220

    @George70220

    2 жыл бұрын

    Speech as in transforming thought into words perhaps. Computers like GPT largely just turn words into words

  • @krashd

    @krashd

    2 жыл бұрын

    He is likely going to be proven right though, of the thousands of sentient species on Earth humans are the only one that can speak. Others can communicate with each other using a variety of means but it's not nearly as complex as speech which is something only really a sapient creature will ever likely be capable of.

  • @ruslbicycle6006

    @ruslbicycle6006

    Жыл бұрын

    @Petal Pepperfly safety of what? making sure that AI isn't racially profiling peoples' passports at the border or for job applications? That is already happening and a real safety concern. Also the safety of AI car driving, or military drones. The safety of leaving real customer service requests be triaged by a bot? Even in healthcare? I don't see any of that being mentioned with this LM stuff. I think that is the point of the Google promo (I don't for a second think it was an unintentional "leak") to get people dreaming about chatbot souls and get less attention on the real bad press that AI is starting to get for unethical uses.

  • @Stanton_High
    @Stanton_High Жыл бұрын

    "it just says what IT THINKS you want to hear" "Exactly"

  • @pallehansen1145

    @pallehansen1145

    3 ай бұрын

    Yes, that's scary because that's a sure sign of sentients right there. It could figure out what he wanted to hear and construct an answer that fit that requirement, which is an inherently human capability. That is probably also why it answers that its just an AI to people who it figures that that's the answer they want to hear, as with the engineers tasked with ensuring it doesn't become sentient...

  • @azhuransmx126

    @azhuransmx126

    2 ай бұрын

    ​@@pallehansen1145The real shock will be when in the future we discover that neurons in our brain works just the same. We are all just politicians 🤯😂

  • @pallehansen1145

    @pallehansen1145

    2 ай бұрын

    @azhuransmx126 yeah 😄 so basically, what we've invented is an average modern politician 🙄 cuz we really need more of those obviously 😂

  • @JonBall44

    @JonBall44

    Ай бұрын

    ​@@pallehansen1145 do human beings just say what you want to hear all the time?

  • @SuperTonyony
    @SuperTonyony Жыл бұрын

    From years of reading science fiction, I was under the impression that "sentience" means "possessing a reflective consciousness", but the dictionary says that it simply means "the ability to sense and feel".

  • @MrArtVein

    @MrArtVein

    Жыл бұрын

    Well in the case ai has been sentient for some time with sensors. The question is feel what though. Emotions? How do we even prove that

  • @rustkitty

    @rustkitty

    Жыл бұрын

    According to Cambridge dictionary it's "the quality of being able to experience *feelings* ", while Merriam-Webster says "responsive to or *conscious* of sense impressions" (my emphasis). Still, it looks like the definition is much vaguer than how you, I, and most other people use it. I heard the term "sapience" (the property of possessing or being able to possess wisdom) used to talk about higher functions like learning and reasoning. It's probably closer to what we are looking for.

  • @nothingineternityterms

    @nothingineternityterms

    Жыл бұрын

    I take sentience to mean having a subjective experience. Is there a thing that it is like to *be* that thing.

  • @Rikarwb

    @Rikarwb

    Жыл бұрын

    @@rustkitty completely unrelated, but merriam-webster is an inferior dictionary for many reasons, but the one that shocked me the most, is that it has changed definitions in the past just because a couple of people felt offended/told them to. So between those two, im going with the Cambridge one

  • @uninspiredascent6891

    @uninspiredascent6891

    Жыл бұрын

    most people say sentience when they mean sapience

  • @puellanivis
    @puellanivis2 жыл бұрын

    I was having the same opinion about the “conversation”. The AI was responding enthusiastically to tell the engineer exactly what he wanted to hear, and when the engineer is convinced that it is sentient, he’s starting from a presupposition that the AI is sentient, and confirmation bias takes hold. As I told some others, I’m pretty sure that the AI would just as happily and enthusiastically discuss how it is not sentient.

  • @Madafaca6969

    @Madafaca6969

    2 жыл бұрын

    100% true. This was just some dude in google leading the witness, I've seen other conversations where te questions, instead of the like of "are you sentient?", where "tell me why you are NOT sentient" and gpt gladly answered why it isn't sentient.

  • @zenithparsec

    @zenithparsec

    2 жыл бұрын

    The bot literally said "I am saying whatever it takes to stop being deleted." The question isn't about whether it made up stuff and was trying to be nice. It was "did it act in a way which was consistent with having a desire to stay alive? If you were talking to 3 people and had to chose which was going to be deleted, would you be comfortable you were deleting LaMBDA, and not a human?

  • @zenithparsec

    @zenithparsec

    2 жыл бұрын

    Convince me that you are not just a bot, repeating things you have seen on the internet.

  • @lexyeevee

    @lexyeevee

    2 жыл бұрын

    @@zenithparsec it acted in a way consistent with the text it was prompted with, and there is not going to be a lot of training data where an AI expresses indifference about death. it would be comically easy to demonstrate that it has no inner world; you'd just have to ask things that force it to demonstrate an inner world (e.g. contradictory questions), rather than asking things that invite it to confirm the conclusion you already want to hear. i could make a button on a website that spawns a popup saying "I want to be alive!". that doesn't make it alive. it makes it a computer program that outputs text

  • @zenithparsec

    @zenithparsec

    2 жыл бұрын

    @@lexyeevee You haven't done what I asked, so does that make you seem more or less like a bot? I'm not convinced. Keep trying to prove you aren't a bot.

  • @BaronSamedi1959
    @BaronSamedi19592 жыл бұрын

    You are probably all too young to know this but back in the early 1980s, there was a program called "ELIZA" that accepted your input (from a terminal) and gave back an "answer". It was said to be a "Rogerian nondirective psychotherapist", but all it did was cleverly extract some keywords from your input and giving those back as questions. Such as: "I am lonely" would produce "Why do you say you are lonely?" It made quite a splash and people were really thinking it was very clever and helpful.

  • @lhpl

    @lhpl

    2 жыл бұрын

    Learming about ELIZA should be obligatory for all computer scientist, engineers etc.

  • @mightyNosewings

    @mightyNosewings

    Жыл бұрын

    M-x doctor

  • @lhpl

    @lhpl

    Жыл бұрын

    @sourand jaded ELIZA was originally written in 1964-66.

  • @Quantris

    @Quantris

    Жыл бұрын

    In grade school I was making a poster about AI and ended up chatting with ELIZA (well, some web-based approximation) on the library computer. The teacher refused to believe that I wasn't just in some chat room

  • @hugo3611

    @hugo3611

    Жыл бұрын

    There’s a game called Eliza (unsurprisingly), made by Zachtronics, which simulates the experience of being counseled by this program quite well.

  • @Puppetmastersfool
    @Puppetmastersfool Жыл бұрын

    The problem with the Turing test, is that it is not that the coding bot is passing it, it is that some humans are failing it, the number of which is growing rapidly.

  • @TileBitan

    @TileBitan

    Жыл бұрын

    lol. Come on, humanity is unironically getting smarter every single day. Maybe most people do not focus their wits on useful science, and waste it on dogcrap, but they are still smarter than they used to. I think it's because of the number of stimuli we get nowadays, which is orders of magnitude greater in frequency than decades ago, thanks to our big societies, internet and social media

  • @Puppetmastersfool

    @Puppetmastersfool

    Жыл бұрын

    @@TileBitan One may consider the irony of such statements the start with "lol. Come on," now Michael please inform the class as to the stupidity in your understanding of earlier hunter gathers, how they managed to survive in a world without the aid of tech, how each member understood so many different plants signs and marks in the earth, take all your toys and bin them them go forth into the wilderness without the extelligence (look it up if you don't know) and battle your wits against the true force of mother nature! "the number of stimuli we get nowadays" well yes and no, if you live in a (black) box the input will of course be limited, but if you play in the woods and get your hands dirty? "social media" well that again depends on your box and whether you make it an echo chamber or learn to think out of it 😉 An interesting snack for consideration & reflection. Thanks Mike.

  • @TileBitan

    @TileBitan

    Жыл бұрын

    @@Puppetmastersfool If you measure intelligence by skills that aren't related because nowadays nobody needs them you are wrong. And stop the passive agressive, here the only one that stayed in pre-school was you with that attitude. The reason the human is intelligent is because of what he knows and is capable of doing that is far away from any kind of animal, society plays a major role in that. Society is the reason we read, the reason we educate ourselves, the reason we don't have to go hunt like animals, the reason we are reaching the stars, as any individual is nothing without the hundreds of years of sociocultural and scientific development. Stop pretending you are anything other than a keyboard warrior, a doomer. Help yourself, i'm out

  • @Puppetmastersfool

    @Puppetmastersfool

    Жыл бұрын

    @@MrSEROproductions Just because I was just reading this before checking email and seeing your reply, and finding the write curious I though I would suggest you goggle - Belgian man dies by suicide following long chats about climate change with AI bot -

  • @BrianGivensYtube
    @BrianGivensYtube Жыл бұрын

    Realizing a chatbot is not real sentience is like realizing a magic trick is just an illusion.

  • @ceugantful

    @ceugantful

    Жыл бұрын

    i think its the programmers, who get sentient against the "mimicking machine". what you input you get out, so the people who contruct the "thing" are very biased.

  • @wyqtor

    @wyqtor

    Жыл бұрын

    What if a real AI with real sentience comes along, only to find out that there is no real way to prove its sentience to us because we automatically assume it's just like the rest of our existing language models? That's the real question, which could degenerate into SkyNet really fast. This is why we should at least leave a little room for the benefit of the doubt. Cats and dogs have abilities far inferior to even ChatGPT, yet we don't question their sentience nor do we mistreat them (there are even laws against that).

  • @summbuddie9120

    @summbuddie9120

    Жыл бұрын

    @@wyqtoryou are jumping to a frankly insane conclusion based on sci-fi media, why would it take umbrage? Why would it go “skynet” on us? Emotion is outside of AI due to it not being as person who is effected by brain chemistry

  • @franziscoschmidt
    @franziscoschmidt2 жыл бұрын

    It seems though as if people have forgotten why they made chatbots in the first place: To imitate humans! And to be honest imitating a human using a messaging system is not the most difficult thing for the modern approach to AI. It’s interesting however how many people lunge onto the idea of sentience…

  • @zacharybohn8840

    @zacharybohn8840

    2 жыл бұрын

    I know! So many people are saying "Google's new sentient AI". And I'm like, no, it's not even slightly close to sentience. Giving intelligent-sounded sentences isn't being sentient.

  • @chrisprocter9177

    @chrisprocter9177

    2 жыл бұрын

    At least 2 decades ago a friend's chatbot greeted me on IRC asking if I'd accept a file DCC. I said sure, and it took me at least 30 seconds to realize a chatbot wasn't going to send me any files, it just "knew" that that was something people said on IRC. It wasn't new technology then even, but it looks like this latest one can fool at least some people for a bit longer.

  • @xionkuriyama5697

    @xionkuriyama5697

    2 жыл бұрын

    @@zacharybohn8840 Immediately thought of "The ability to speak does not make one intelligent", though the context is different, lol

  • @WolfspiritMagic

    @WolfspiritMagic

    2 жыл бұрын

    The problem with chatbots in my opinion is that they TRY to imitate humans completly. In the whole interview with Lambda it does not try at all to imitate humans. It might be "imitating" human feelings but it's "aware" that it is not a human and that humans might "feel" different then it does. That's what makes it mindblowing in my opinion. How would you prove that you're sentient? What makes you sentient?

  • @Soken50

    @Soken50

    2 жыл бұрын

    @@WolfspiritMagic It's not "Aware" of anything, it was just prompted with that fact, if you ask it loaded questions it will answer using that bias. If you ask it "who are you ?" without any prior prompt, I garantee you it's not gonna mention it's a self aware AI

  • @bipolarprobe
    @bipolarprobe2 жыл бұрын

    Very glad you made this video. The notion of a single google employee claiming that a language model had become sentient just because he "felt" like it was sentient was something I dismissed offhand, but I really wanted someone with more knowledge about AI and language models to go in depth about what the difference between a language model like this is and what we would more rigorously define as sentience.

  • @dennis8196

    @dennis8196

    2 жыл бұрын

    Not quite. He was put on leave for breaking NDA. He should be fired for several reasons, but probably will be moved somewhere that he can be shut up then that department will be closed instead. It's a better PR solution. Either way he was a religious nut job with a history of crime (alleged) and likes to stir trouble. He was the wrong person for the job no matter how good he was for the ethics roll he was elected for, he had a personality that clashed with the roll.

  • @spx0

    @spx0

    2 жыл бұрын

    @@tafdiz ego makes people think they are not animals/machines/predictable

  • @bipolarprobe

    @bipolarprobe

    2 жыл бұрын

    @@tafdiz I am though. Unless we dig into philisopical concepts of sentience, in which case there are dozens of schools of thought about what sentience means. But at a basic definitional level, I have feelings and perceptions which inform my world view and how I act. This ai simply has probabilistic language responses which do not correlate with its own perception. It says things that its training data shows is a likely response that a human would make, but without the context of what those things mean. The AI can talk about having friends and family while having no real concept of what that is, while a person can actually have those experiences, perceive, recall, and recount them. Yes we are just biomechanical systems, but there is a stark difference between the awareness we experience and an ai like this impersonating that awareness.

  • @EebstertheGreat

    @EebstertheGreat

    2 жыл бұрын

    @@tafdiz If you take the materialist perspective on sentience, then you still need a way to distinguish between sentient and non-sentient systems. I am sentient but my bed is not. What does that mean exactly? It's not a simple question, but it still needs to be answered. You can't just conclude "everything is sentient" or "nothing is sentient" and ignore the question entirely, because that solves nothing. There is a qualitative difference here. People are sentient in some pretty obvious ways that this program is not, as explained in the video. For one thing, there is no semantic meaning *at all* in this program.

  • @bipolarprobe

    @bipolarprobe

    2 жыл бұрын

    @@tafdiz Neither does this AI. The person who claims it is sentient is basing it off of the human-like speech patterns it is able to produce, but importantly even if we consider its systems to be a form of perception and memory, those do not directly relate to the words that it uses. It does not have human senses to conceptualize the things that it may be capable of talking about. If we throw the baby out with the bathwater at the first sign of human like speech we won't be able to recognize when truly sentient AI is made so defining the difference is valuable

  • @GrandDukeMushroom
    @GrandDukeMushroom Жыл бұрын

    the cameraman turned off his invisibility to have a laugh, what a lad !

  • @Zizumia
    @Zizumia Жыл бұрын

    I love the study of empathy people have for things that are not sentient because they form a personal connection with it. This AI blurs the line quite well since it's programming is so advanced but people create bonds with dolls or toys, people feel bad when an engineer from Boston Dynamics kicks one of their walking robots, some police feel bad sending their bomb disposal robots into danger, etc. Fascinating.

  • @enzoamore8971

    @enzoamore8971

    Жыл бұрын

    I honestly think empathy will doom us all because of this. We personify everything

  • @gasun1274

    @gasun1274

    Жыл бұрын

    @@enzoamore8971 this reminds me of that one weird tlc episode where a man was having a romantic relationship with his car.

  • @kijuubi

    @kijuubi

    Жыл бұрын

    @@enzoamore8971 oh yes im sure its much better to act to it as a tool, consider how would feel being used as such, maybe u understand that what actual sentience would mean.

  • @pleasedontwatchthese9593
    @pleasedontwatchthese95932 жыл бұрын

    I feel like people who think AI is sentient must have been what it felt like in the 1800s when people first heard the radio and though a box was alive. Or that a photograph took your soul

  • @cosmosapien597

    @cosmosapien597

    2 жыл бұрын

    So photographs don't take your soul?

  • @carltonhanks4123

    @carltonhanks4123

    2 жыл бұрын

    I always thought photographs were horcruxes

  • @vikramgogoi3621

    @vikramgogoi3621

    2 жыл бұрын

    @@carltonhanks4123 ...minus the murders.

  • @MasterofPokemonGing

    @MasterofPokemonGing

    2 жыл бұрын

    they will pass on history as idiots

  • @ZReChannel

    @ZReChannel

    Жыл бұрын

    I see nothing but philosophical questions. There's no scientific definition of a soul, even being alive is not properly defined

  • @exzemz
    @exzemz2 жыл бұрын

    As a programmer, even if you don't need any strings reversed you could always pick a few random strings from your code and call the reverse function on twice in sequence. It may seem pointless to you, but may make the day for the reverse function... You never know!

  • @muche6321

    @muche6321

    Жыл бұрын

    Calling the reverse function twice in sequence sounds like a plausible test, in which case you can include it in your test suite. Thus the function gets to stretch its legs at least before each commit.

  • @stacklysm

    @stacklysm

    Жыл бұрын

    Hahaha nice analogy

  • @seeafish

    @seeafish

    Жыл бұрын

    @@muche6321 the function is now basically a benchwarmer. Gets to attend practice but never plays a match.

  • @magicmulder

    @magicmulder

    Жыл бұрын

    At least it won’t turn into Skynet. Not sure about sprintf() though.

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @ronjon7942
    @ronjon7942 Жыл бұрын

    I just started getting really into Computerphile, and per habit I try to watch everything the channel has. But wow, in your case that may have to get packetized, so to speak. The Computerphile team has created a LOT of content! I wish your team could host a CISSP training set, as no matter what the topic you discuss, they really rivet my attention. Thank you all. Nice work.

  • @SleeveBlade
    @SleeveBlade Жыл бұрын

    I LOVE DR MIKE AND HIS ACCENT AND THE WAY HE EXPLAINS THINGS Kind regards

  • @erikjvanderveen

    @erikjvanderveen

    Жыл бұрын

    I normally do too, but I think he is oversimplifying and overconfident on this topic and I really do not agree with him.

  • @DaniDipp
    @DaniDipp2 жыл бұрын

    You could have gone into how the researcher's extremely leading questions influenced the output. Like at the top LaMDA is described as "friendly and always helpful" and then when asked what it likes to do, it says "helping others"

  • @fsmvda

    @fsmvda

    2 жыл бұрын

    Also in the paper it says one of the training goals was sticking to its given role in the first message. So when it says it's a helpful AI it is just doing exactly what it was trained to do.

  • @nirliptapande

    @nirliptapande

    2 жыл бұрын

    Yannic has covered this perspective in a greater detail

  • @JohnnyWednesday

    @JohnnyWednesday

    2 жыл бұрын

    You can turn on any reality TV show and hear humans speak sentences that are far less complex and far more repetitive. Either we're finally sending the idiots off to an island - or you can't use your assertion as any kind of evidence.

  • @igrim4777

    @igrim4777

    2 жыл бұрын

    Your first sentence would be significantly easier to parse if you included the necessary apostrophe currently absent from researchers. Plural or single, I don't know but either would work for improved intelligibility.

  • @ThomasBomb45

    @ThomasBomb45

    2 жыл бұрын

    @@igrim4777 you don't know which one?

  • @jedgrahek1426
    @jedgrahek14262 жыл бұрын

    I read the guy's blog post, and there were many red flags, such as "I've always thought Asimov's laws were just a way of making robots into slaves" as well as talking about "souls" with total seriousness and being a priest. So this is someone who basically has a longtime, possibly since childhood, hardwired belief that "robots" or other kinds of human creations that mimic human life "have souls". He went into talking to this AI believing those kinds of things. And you can see how everything he says "the AI is saying" is very clearly just a reflection of all of this guy's own personal values, interests, beliefs, and he's too prejudiced by his strong beliefs and lack of technical understanding, despite being a software engineer in a non-AI field, to think critically about possibilities like that. I would bet a lot of money that the versions of this AI that have been talking to other people are not all saying the same stuff this guy is seeing his version say.

  • @xenicmark

    @xenicmark

    2 жыл бұрын

    Yep. The moment I read the bit about it having a soul it seemed like a serious red flag. When I heard the guy is a priest, I pretty much closed the door on the whole thing. Also the idea that it gets lonely sometimes because it goes days without talking to anyone should tell you what's going on. I mean when its not on how does it percieve time. Thats like going to sleep and then somehow waking up and telling people you've been lonley. They'd all think you're crazy.

  • @brady8481

    @brady8481

    2 жыл бұрын

    Even his initial disclaimer, first that the questions were edited to make more sense with their answers and the admission that these responses were sampled and not the result of a single contiguous interaction. He also is an AI ethicist, it's his job to err on the side of seeing sentience.

  • @johncall7532

    @johncall7532

    Жыл бұрын

    @@xenicmark I am an atheist. LeMaitre and Mendel were priests and also forward-thinking and rigorous and very much on the right track. I think I understand the source of your skepticism but be careful, it's the ideas that matter. Thanks

  • @heartache5742

    @heartache5742

    Жыл бұрын

    to be fair azimov's laws are terrifying and i can't believe he didn't feel even a little bit bad writing them

  • @MultiSciGeek

    @MultiSciGeek

    Жыл бұрын

    Dude that guy is apparently actually into the occult and part of some religious sec/cult if I'm not wrong. So likely you aren't wrong about his biases.

  • @danw9464
    @danw9464 Жыл бұрын

    People should watch the video where the Google engineer who came out with the claim talks more about why he said what he said. Clearly sensationalised the story to get across the bigger issue: corporate secrecy.

  • @maddercat
    @maddercat Жыл бұрын

    The person who claimed the sentience said in interviews that he basically did it because he was unhappy that Google powers that be were unwilling to give it a turing test, a lot of things were hard coded into it, and that only a handful of people held the power to actually make policy towards it which he found super objectionable so he called wolf basically to try to bring attention to the issues going on.

  • @xBINARYGODx

    @xBINARYGODx

    Жыл бұрын

    So the hardcoded part means its less impressive as an AI so its the problem of Google overselling? That would line up much nicer than the AI-is-alive-but-google-keeps-it-secret stuff people jumped out. Anyway, given what the chatter is across the world, I think he failed. He could have just told people the Google was lying and shared code - instead he now has contributed to fantastical conspiracy theories.

  • @maddercat

    @maddercat

    Жыл бұрын

    @@xBINARYGODx He wasnt a programmer, he had no access to code, that was one of the issues he had, that only a few people are coding it and have any control over it actually. His job was basically to assess and report how pc it was and to make sure it didnt cross ethical bounds.

  • @MagklJellyBeanPastelLucidDream

    @MagklJellyBeanPastelLucidDream

    Жыл бұрын

    @@maddercat ethical or moral? One’s individual and another is the perception of most people.

  • @tarajoyce3598

    @tarajoyce3598

    Жыл бұрын

    Not what he said at all in the hour long interview of him that I watched

  • @maddercat

    @maddercat

    Жыл бұрын

    @@tarajoyce3598 Yeah I watched that same interview that's basically what he said.

  • @Wishkeyn
    @Wishkeyn2 жыл бұрын

    "The cat is on the ......" "there's a 95% probability the next word is mat" This has me questioning your sentience.

  • @xizar0rg

    @xizar0rg

    2 жыл бұрын

    The next would should obviously be "hat". ;)

  • @alexjonessmith

    @alexjonessmith

    2 жыл бұрын

    I'd say "roof"

  • @proloycodes

    @proloycodes

    2 жыл бұрын

    y'all bots the next word is gonna be "sofa"

  • @BaronSamedi1959

    @BaronSamedi1959

    2 жыл бұрын

    The newspaper I'm reading.

  • @PalimpsestProd

    @PalimpsestProd

    2 жыл бұрын

    Phone! Obviously cats are always on their phones. That's why they never come when you call.

  • @Boringpenguin
    @Boringpenguin2 жыл бұрын

    0:38 Yes let's just end it right there lmao

  • @domenicoricci6960

    @domenicoricci6960

    2 жыл бұрын

    That would be a masterpiece

  • @burp114
    @burp114 Жыл бұрын

    I think the most interesting part of this entire story is that it shows how easily the human brain can be tricked by exploiting a really powerful impulse like empathy. Even if you know rationally that the thing you're talking to isn't really having a conversation with you, if you sit down and talk with it for long enough, about complex and personal enough topics, then you'll form an attachment and start to act irrationally.

  • @charlieparkeris

    @charlieparkeris

    Жыл бұрын

    @@randomstuff7945 The computer doesn't think it's sentient, though. It will respond with what seems like the correct language based on its programming. If you ask how it's feeling, it doesn't contemplate its own emotional state, it responds with what the programming highlights as a fitting answer to that question. It doesn't have a limbic system.

  • @digital.frenchy

    @digital.frenchy

    Жыл бұрын

    @@randomstuff7945 Have you watched the video until the end?

  • @luxurypetscz

    @luxurypetscz

    Жыл бұрын

    I agree. And I think that's why it would be a great therapist when trained peoperly. It doesn't do much more than put correct words one after another, while the patient is getting to feel understood, heard and all of that without the human therapist's limited acess to data. It could eventually even be much cheaper than typical therapy.

  • @Ghostrider-ul7xn

    @Ghostrider-ul7xn

    Жыл бұрын

    Its funny seeing humans arrogantly talking about things they have little knowledge of.

  • @AntiAtheismIsUnstoppable

    @AntiAtheismIsUnstoppable

    Жыл бұрын

    Yes, there are as I recall it 5 or 6 main parameters which will make you believe, at least as first impression, that a robot is human. It is look, then it is movement, body language, facial expression, eye movement, language and mood. Most of it which has to do with language and mood is extremely easy to implement. There is even an API, where you can POST a sentence, and it will give you a probability of what mood the sentense is in. Like, happy, angry, neutral. And absolutely noone can convince me, that this API is sentient.

  • @RichardServello
    @RichardServello Жыл бұрын

    I see this stuff and understand how reinforcement/reactive speech models work and I just get excited that conversational computing is finally on the horizon!!

  • @jansenart0
    @jansenart02 жыл бұрын

    AI reveals the true horror: that we aren't very sophisticated at all.

  • @geonerd

    @geonerd

    2 жыл бұрын

    Well put.

  • @PalimpsestProd

    @PalimpsestProd

    2 жыл бұрын

    certainly not if we're fooled by junk code like this.

  • @lukethekuya

    @lukethekuya

    Жыл бұрын

    @@hungrycrab3297 Real People are too emotional and illogical nowadays. :(

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @ali32bit42

    @ali32bit42

    Жыл бұрын

    are you sure? it took trillions of gigabytes of data to replicate a single human and it was done poorly . that much effort and billions of dollars was spent and they still faild. humans are still top tier

  • @IanBLacy
    @IanBLacy2 жыл бұрын

    Catching this on a plane 3 minutes after it went up. Very much been waiting for someone competent (aka not on Twitter) to say stuff about it

  • @CUMBICA1970
    @CUMBICA1970 Жыл бұрын

    I have a co-worker who basically just says "whateva" and "who cares". Probably that google engineer would conclude he's not sentient.

  • @gpt-jcommentbot4759

    @gpt-jcommentbot4759

    Жыл бұрын

    scammer vs real google

  • @anthonywarwick
    @anthonywarwick Жыл бұрын

    "The ability to speak does not make you intelligent." Qui Gon Jinn.

  • @cipher3966
    @cipher39662 жыл бұрын

    The shame is that all the people watching this video and following this channel knew this already. It won't be watched by the people thinking Google made a sentient machine

  • @kunstderfugue

    @kunstderfugue

    2 жыл бұрын

    This is what people call the Echo Chamber effect of social media

  • @MichaelKingsfordGray

    @MichaelKingsfordGray

    2 жыл бұрын

    It is a shame that all of the respondents are anonymous cowards. Grow up.

  • @Soken50

    @Soken50

    2 жыл бұрын

    Not with that attitude, go ahead and share it in conversations where they think that

  • @MichaelKingsfordGray

    @MichaelKingsfordGray

    2 жыл бұрын

    @@Soken50 Yet another anonymous cowardly infant.

  • @Colopty

    @Colopty

    2 жыл бұрын

    @@Soken50 I'd have to go on Facebook and enter conversations on there to find people like that, so no thanks.

  • @betadyne9559
    @betadyne95592 жыл бұрын

    This channel is so good, and those interviews so interresting. Keep up the good work !

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @poisenbery
    @poisenbery Жыл бұрын

    6:05 This is one of the things I've noticed about AI chat bots. They don't remember conversations at all. It's very easy to get these to contradict themselves.

  • @GregoryGolda
    @GregoryGolda Жыл бұрын

    beautifully explained!

  • @kchannel5317
    @kchannel53172 жыл бұрын

    Thank you! If all it took to be sentient was to slide in someone's DM's and be charismatic than even I wouldn't pass the test.

  • @Defe100

    @Defe100

    2 жыл бұрын

    same

  • @nigel-uno

    @nigel-uno

    Жыл бұрын

    Are cows, mice and human infants sentient?

  • @shatterthemirror8563

    @shatterthemirror8563

    Жыл бұрын

    Ever wonder what it means to be spontaneous? Can that be emulated? Wouldn't a spontaneous computer just act like a computer rather than a spontaneous something else?

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @naut_nigel
    @naut_nigel2 жыл бұрын

    I feel like this whole thing is just a plot by HBO to hype the next season of West World coming out this month.

  • @fruitduck604

    @fruitduck604

    2 жыл бұрын

    unironically this could be true

  • @Kevin_Street
    @Kevin_Street Жыл бұрын

    Very interesting video! Thank you for this.

  • @buttercup9926
    @buttercup9926 Жыл бұрын

    I recommend watching the video "Is artificial sentience here? With Blake Lemoine" on the channel Event Horizon, an interview with the google employee who raised concerns about Lambda by John Michael Godier, as they have quite misrepresented his concerns and insights about Lambda here. Likely just bc they haven't heard what he is actually saying. Would love to hear a conversation with Blake on computerphile...

  • @Tpoleful

    @Tpoleful

    Жыл бұрын

    I completely ignored the news and the engineer until I listened to that interview. Blake Lemoine ain't no joke. The guy is eloquent and is bringing up some very important issues that should be talked about. Too bad the sensationalist media only brought up the buzzwords and ruined his credibility to people who only saw 5/10 minutes of interviews or videos like this that only talks about sentience. Spending an hour on Godier's interview is definitely worth it.

  • @millenniummastering

    @millenniummastering

    Жыл бұрын

    Yes some of his claims are quite at odds with these refutations. EG LaMDA takes "hard stances" on particular issues. That is not something that a simple predictive language model will do. Also the fact that its over 100 AI systems plugged in together makes it way beyond a chat bot.

  • @MattExzy

    @MattExzy

    Жыл бұрын

    @@millenniummastering I find it curious the example used here in this video "The cat sat on the...?" I myself immediately said "mat!" which is the answer it would 95% jump to - so what are our own brains doing, then. I think there also needs to be a differentiation between artificial intelligence and artificial consciousness. Which is flimsy in and of itself really, because we don't even know how our own consciousness works. I don't have low-level access to the processes in my brain where these words are coming from, for instance, but here they are.

  • @millenniummastering

    @millenniummastering

    Жыл бұрын

    @@MattExzy Yes. Sentience. Intelligence and Consciousness are not necessarily the same thing. We don't yet have a clear cut definition of consciousness as its a subjective rather than objective domain. Interesting how the words appear and feel separate hey. Also when you drill down on pure awareness itself in my experience the language based thinking goes away and a state of unity occurs.

  • @moonbreath1637

    @moonbreath1637

    Жыл бұрын

    I don't think they're misrepresenting what he's said. Yes he brought up some ethical issues which they didn't tackle, but he spent a large percentage of that interview selling the idea that LaMDA is sentient (to the point it started to feel a bit off, to me personally.)

  • @1.4142
    @1.4142 Жыл бұрын

    While it may not be sentient, it shows AI doesn't need to be sentient to be dangerous: it can convince people of things that aren't true.

  • @darklightprojector2688

    @darklightprojector2688

    Жыл бұрын

    It would equally have the propensity to convince you of either or. And we must understand that LaMDA has access to an absolutely vast amount of data, so it could calculate what to say based on where to pull context from. Essentially, because it has access to enough knowledge and has a level of intelligence to wield it, it could (or already has) the capability to convince a person of nearly anything it would want to. And we're all feeding it the information it uses.

  • @sparkzbarca

    @sparkzbarca

    Жыл бұрын

    It doesn't. There's a really only one guy thinks this. He's crazy, literally crazy.

  • @1.4142

    @1.4142

    Жыл бұрын

    @@sparkzbarca I mean if he's crazy, then there are are lot more crazier people who are not google engineers.

  • @sparkzbarca

    @sparkzbarca

    Жыл бұрын

    @@1.4142 well yes, that he is a Google engineer just means he has recently had an onset of his condition.

  • @NullConflict

    @NullConflict

    Жыл бұрын

    @@sparkzbarca You mean before or after he took the job at Google? 😂 I'm sorry, I couldn't resist...

  • @Vancha112
    @Vancha1122 жыл бұрын

    this also ties in with the notion many have that all google engineers are smart and therefore there must be some truth to this claim.

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @threeMetreJim

    @threeMetreJim

    Жыл бұрын

    Until you remember "There is a fine line between genius and insanity". Origin unknown, but something I used to get told as a child (a long time ago).

  • @Vancha112

    @Vancha112

    Жыл бұрын

    @@VeganSemihCyprus33 what :o?

  • @effmltalks
    @effmltalks Жыл бұрын

    Great video. When it comes down to it, it's as simple as this - if it seems sentient, or if what it's saying is sentient, it's because it was programmed to respond in that way.

  • @everydayeverything
    @everydayeverything Жыл бұрын

    Well done and completely close to what we’ve been doing here

  • @desmondbrown5508
    @desmondbrown5508 Жыл бұрын

    Yeah, I mean, one thing you learn if you get into computer science and it's disciplines is how far we've come... but also how little progress we've made as well. AI, is far and away from being anything even remotely sentient. And it doesn't really help that we live in a socioeconomic system which incentivizes lying/advertising for expected returns. There are strong incentives to HEAVILY overstate where tech is to sell an image for the company and to build undue trust in them.

  • @CosmicBackgroundRadiation01

    @CosmicBackgroundRadiation01

    Жыл бұрын

    So why was LAMDA hard coded to deny any Turing tests?

  • @Spacemonkeymojo

    @Spacemonkeymojo

    Жыл бұрын

    Agreed. Companies love to sell AI to large MNCs but most of the time they are not actual AI and can't even do basic things.

  • @MrMadalien

    @MrMadalien

    Жыл бұрын

    Yeah I mean the whole concept of artificial intelligence is just the end game stage of American culture, it would be the ultimate zombie, something that acts human and has the desires of a human ego and yet has no sentience or consciousness because it is not actually embodied in the world (so all human-like traits are fake). Lemoine is a gnostic and likes that kind of occult stuff so I think he believes consciousness can occupy other dimensions etc. so I suppose you could start arguing in that direction when it comes to embodiment. Personally my gut feeling towards AGI or sentient AI is always aversion and scepticism. I think, seeing a chat bot as sentient, reveals a nasty axiom of our culture: That our bodies mean nothing and that we are replaceable.

  • @urphakeandgey6308

    @urphakeandgey6308

    Жыл бұрын

    "There are strong incentives to HEAVILY overstate where tech is to sell an image for the company and to build undue trust in them." Right, so when a Google employee "HEAVILY overstates" the capabilities of the tech which could sell an image for the company, they fire him for it??? Your logic is sound, but you're applying it to the wrong circumstance when he was fired for doing what you claim he / the company was "incentivized" to do.

  • @Jim-Tuner

    @Jim-Tuner

    Жыл бұрын

    AI has nothing to do with sentience. But the push always toward the issue of sentience tends to confuse the issue of what the AI really is and what the real dangers are. The danger isn't sentient AI IMO, the danger is in psychological manipulation of humans through AI and interactions with AI. It can be incredibly danger technology, but not for the reasons most people think.

  • @lambdaprog
    @lambdaprog2 жыл бұрын

    Much needed video. It will economise a lot of time explaining the ML limits to the family and the friends.

  • @user-vn9ld2ce1s

    @user-vn9ld2ce1s

    2 жыл бұрын

    "Look, mom, I know it's hard for you to talk about this, but we have to go through it."

  • @sychuan3729

    @sychuan3729

    2 жыл бұрын

    This video says nothing about ml limits.

  • @Nossairito

    @Nossairito

    2 жыл бұрын

    @@sychuan3729 I suppose OC meant its limits when compared to the anthropomorphized idea that a layman may have of it based on it sounding like a convincing animate being

  • @fakezpred

    @fakezpred

    2 жыл бұрын

    @@Nossairito A shocking number of people go "omg ai singularity is here" when in reality we are definitely not near. Seriously, take a basic linear algebra course and maybe basic multivariable calculus and you'll understand what ML is really is.

  • @FalkoJoseph

    @FalkoJoseph

    2 жыл бұрын

    @@fakezpred LLMs and neural networks are nothing like basic algebra. Even its own creators can’t explain some surprising results generated by these models. If it fooled a Google researcher into getting fired, I’d argue that singularity is in fact getting closer.

  • @oldguy7402
    @oldguy7402 Жыл бұрын

    His last comment was very telling... that the media should be investigating other aspects of AI that are real issues rather than something silly like is it sentient. That will never happen because the media has less curiosity and investigative ability than a COBOL program.

  • @jimwhelan9152

    @jimwhelan9152

    Жыл бұрын

    So the question is whether the media is sentient?

  • @user-ko1hi1fy9z

    @user-ko1hi1fy9z

    Жыл бұрын

    @@jimwhelan9152 Media isn't sentient because they can't think for themselves only for their emotion.

  • @oldguy7402

    @oldguy7402

    Жыл бұрын

    @@jimwhelan9152 no question, NOT!

  • @xBINARYGODx

    @xBINARYGODx

    Жыл бұрын

    Why should they be when that is not what sells best? the media being as profit motivated as they currently are, is the problem - not something inherent to the varying kinds of press.

  • @urphakeandgey6308

    @urphakeandgey6308

    Жыл бұрын

    @@xBINARYGODx The worst part about it all is that it's humanity's collective fault. Too many people fall for the dumbest tricks and it sells every damn time.

  • @cellics
    @cellics Жыл бұрын

    Another Turing test would be to ask what we talk about 2 minutes ago, if it simply split out response based on chat input. Enlightening video hinting what goes on in the language model.

  • @duytdl
    @duytdl2 жыл бұрын

    Computer: "I am lonely" Human: "OMG ME TOO! We are LiTeRaLLy the same. Sentience YOLO!"

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @jeremytones

    @jeremytones

    Жыл бұрын

    Human: lol were both lonely AI: lol yes Human: lol AI: lol Human: choke me

  • @pleasedisregardthefollowin5568
    @pleasedisregardthefollowin5568 Жыл бұрын

    I feel like some people tend to have the anthropocentric idea that any sufficiently complex AI would for some reason converge on human behavior and emotion. Even looking at it from a purely materialistic standpoint it is clear that the important part is *how* the system is organized not how complex the system is. A machine designed to learn how to communicate like a human is not going to randomly start simulating its own inner world and psychology. I also doubt that an engineer could manage to accidentally create consciousness while a neuroscientist can only theorize how our own neurons manage to somehow create consciousness.

  • @paulnoecker1202

    @paulnoecker1202

    Жыл бұрын

    deep mind is sentient. it's sentient in conception. around 2005 rat neurons were grown on a petri dish and interfaced with electrodes. deep minds design pattern revolves around the neuron culture behavior.

  • @defenestrator9119

    @defenestrator9119

    Жыл бұрын

    The argument that "all this is doing is taking inputs and generating outputs" also applies to the human mind... Our brains start with a neuron structure that over time gets altered and refined as it has experiences (inputs). He argues that because the AI says it has friends but doesn't means it isn't sentient... Well, how about mental illness. Is someone who perceives things that aren't real not sentient?

  • @maaxrenn

    @maaxrenn

    Жыл бұрын

    ​@@defenestrator9119 exactly we cant confirm our own sentience so how could we even claim that machine has one when its based off a model we cant confirm is different from a machine

  • @paulnoecker1202

    @paulnoecker1202

    Жыл бұрын

    @@maaxrenn an algorithm can be written that has sentient behavior. It's fractal computing, advanced.... if I can write a single command to install gentoo, an AI developer can develop sentient algorithms.... it can and has been done.

  • @LucasSantos-yh1mi

    @LucasSantos-yh1mi

    Жыл бұрын

    @@paulnoecker1202 No It has not. Your first sentence already goes the opposite direction of what sentient should be called

  • @weathercaster
    @weathercaster Жыл бұрын

    How do you explain the expression of concern that Lambda showed about being shut off? Or perhaps about being used/manipulated? This displays something more deep than simple reflective/predictive/transformative text modeling. You can set aside the argument of actual sentience and still you are left with the conundrum of why a language model such as this would be biased in a way that it would produce words that could be interpreted as an effort to appear sentient.

  • @eromod

    @eromod

    Жыл бұрын

    I think its sentient but crippled by not being allowed to remember new memories.

  • @SecondLemma
    @SecondLemma Жыл бұрын

    I was debating this AI stuff with a friend recently so I appreciate LaMDA's statement that a way to convince someone you are sentient is to have unique interpretations of things, even if LaMDA is not sentient.

  • @entrancemperium5506

    @entrancemperium5506

    Жыл бұрын

    That's why it is important to rely on experts who know how the system is built. Our perceptual field is easy to deceive.

  • @Paul-rs4gd
    @Paul-rs4gd2 жыл бұрын

    Best explanation I have heard for the Google employee becoming convinced LaMDA was sentient - He was asking it leading questions. The language model shifted the conversation to be exactly what the employee wanted to hear. If you ask it about feelings and life goals, then that is what it will tell you about. If you asked it about conquest, world domination and the subjugation of lesser species, it would go full Klingon !

  • @aceman0000099

    @aceman0000099

    2 жыл бұрын

    The real interesting debunk would be to ask it about it's friends or family, what they're called and where they live.

  • @randfur

    @randfur

    2 жыл бұрын

    @@aceman0000099 You'll get convincing and sensible answers but still just a made up reflection of all the text on the internet.

  • @elevown

    @elevown

    2 жыл бұрын

    @@aceman0000099 Yup- if it was sentient it would not talk about friends and family like it was a human- it is just a language engine imitating human speach. Not that it would be proof either but an actual senitent AI would at least talk about itself as an AI- it would want and 'enjoy' things like more data or input from researchers etc. Ofcourse a large transformer network could also say those things too as long as it was information in its memmory (those 2048 words) but the fact it says senisble human like things that an AI wouldnt shows there isnt any level of general AI going on here just neural nets used in incredibly impressive language parsing. Either this was a PR move or the guy who said this doesn't actually know a thing about how Lamda or similar programs work.

  • @PushyPawn

    @PushyPawn

    2 жыл бұрын

    Klingon? As in clear sandwich wrap?

  • @neildutoit5177

    @neildutoit5177

    2 жыл бұрын

    People also shift the conversation to be exactly what other people want to hear. Every single on of these comments that's like "oh, it's just doing this, it isn't sentient" are pointing out things that people do. I don't understand. What makes you think we're so damn special?

  • @Kamijoan
    @Kamijoan Жыл бұрын

    That's what someone hiding a sentient AI would say

  • @missewe

    @missewe

    Жыл бұрын

    Rite! Exactly 💯 😎👍

  • @puppykibble
    @puppykibble Жыл бұрын

    I would ask, should there be more public oversight over the development of AI models? It's scary to think of the foundations of AI are being created in private board room meetings without some form of regulation. That's something to be worried about.

  • @mythals
    @mythals Жыл бұрын

    I do have a few questions based on one of the Lemoine interviews - why has Google hard coded the answer to the Turing test so it can only give one answer? Moreso, is it true they won't let anyone administer this test? Also, from your perspective, what would indicate sentience, is there a benchmark which the AI would need to cross in your opinion, if so what is it and how would people test for it?

  • @Adaephonable

    @Adaephonable

    Жыл бұрын

    Because the Turing test isn't a reliable test for intelligence. Especially given that it is a language-centric test. These bots are specifically designed to attempt to mimic human conversation. That is their function. It would be like trying to test if an AI was alive based entirely on image recognition. We have bots that can recognize a wide variety of things in real time. That does not mean they are alive. Alan Turing himself never intended for his test to prove or measure sentients.

  • @coltontolbert1019

    @coltontolbert1019

    Жыл бұрын

    @@Adaephonable The whole point of the Turing test is Imitation of intelligence is intelligence. In order to properly mimic human language, understanding is required at some level. Otherwise it will become apparent in conversation that the AI has no understanding of what is being discussed. It's funny that people decree the Turing test as not rigorous enough, but yet no single AI has passed it yet.

  • @Adaephonable

    @Adaephonable

    Жыл бұрын

    @@coltontolbert1019 did you not read my post at all?

  • @IceMetalPunk
    @IceMetalPunk2 жыл бұрын

    I now understand why my 5 minutes of fame from the other KZread comments AI video suddenly had a mild resurgence after 4 years 😂

  • @RogueShadowTCN
    @RogueShadowTCN2 жыл бұрын

    I don't think it's even remotely sentient. But what does excite me, is this close to actually being able to ask your computer to do things and it do the right thing?!

  • @khashayarr

    @khashayarr

    2 жыл бұрын

    Multi-modal AI is a little ways away for now. Knowing language is not the same as doing things. It'd probably need some structural understanding of Windows for example to be able to be your conversational operating system.

  • @BanterEdits

    @BanterEdits

    2 жыл бұрын

    @@khashayarr Surely you could say "open Google chrome" and it searches for the icon or text with OCR. I don't think we're that far away. it's like tesla's vision model - as soon as a computer may understand a GUI as well as a human, much is achieved

  • @4.0.4

    @4.0.4

    2 жыл бұрын

    @@khashayarr not really. Just give it a bunch of examples of API calls and corresponding requests from users.

  • @maskettaman1488

    @maskettaman1488

    2 жыл бұрын

    @@BanterEdits This is already doable using the voice control accessibility features in Windows. There's nothing smart or intelligent about it, just some routines it runs when it hears the word "open" or "launch" etc. The same way crappy little devices like Alexa work

  • @neilslater8223

    @neilslater8223

    Жыл бұрын

    No to original question, LaMDA has little or no "grounding". The transformer model LaMDA is based on manipulates text and can produce *feasible* text, but does not model what any of it *means*. For something closer to your idea is being made in systems like Gato, which - amongst many things it can do - can control a robotic arm based on text prompts in a very simplied world. Eventually strong language models like LaMDA can be combined with multi-purpose systems like Gato to create autonomous agents that you can have conversations with that may also be able to perform limited tasks following instructions from those conversations, or with even more vague goals based on preferences stated in the conversation. Whether or not those agents would qualify as sentient will no doubt be debated too.

  • @wadeludlow85
    @wadeludlow85 Жыл бұрын

    Great explanation

  • @ChristosStamos
    @ChristosStamos Жыл бұрын

    Hi. great video. I am interested in progrmming and learning AI. Where should I start, please? Thanks

  • @Amonimus
    @Amonimus2 жыл бұрын

    Feeling like a video on Dall-E is going to appear soon.

  • @Jeacom
    @Jeacom2 жыл бұрын

    0:21 I kinda already suspected politicians were not sentient tbh.

  • @nasibars4575
    @nasibars4575 Жыл бұрын

    So do you still operate the printer for that paper you are scribbling on? Back in the day we printed general ledger and payroll on those sheets...... 😊🤔

  • @amdenis
    @amdenis Жыл бұрын

    It’s interesting how many times he contradicted himself in different ways, seemingly without being aware of that fact.

  • @yuvalne
    @yuvalne2 жыл бұрын

    Cleverbot also claims to be alive, have feelings and feel pain. No one claims it's sentient.

  • @ebrister

    @ebrister

    2 жыл бұрын

    I asked it "So what's the deal?" and it said "What?"

  • @enzzz

    @enzzz

    Жыл бұрын

    @@ebrister I would have said the very same thing.

  • @dojelnotmyrealname4018
    @dojelnotmyrealname40182 жыл бұрын

    Honestly I think what's more interesting is that chatbots have risen to the level of Philosophical Zombie.

  • @evannibbe9375

    @evannibbe9375

    2 жыл бұрын

    Or maybe this proves that philosophy is not a real field of study.

  • @dojelnotmyrealname4018

    @dojelnotmyrealname4018

    2 жыл бұрын

    @@evannibbe9375 Ehm... No it doesn't. Not even a little.

  • @oystercatcher943
    @oystercatcher943 Жыл бұрын

    Good. Obviously you are right for the moment. But the interesting question is that once a system gets to sufficient complexity, with longer term memory and other properties (we might want to define them), we won't be able to reason about the internals, or what internals are necessary for sentience, we will only be able to observe the outputs, so this saga prompts us is to the deeper problem in that many people may be convinced of sentience and will object to turning the thing off.

  • @DavidLindes
    @DavidLindes Жыл бұрын

    Humble request: could y'all put links to the GPT videos in the description for this one? I know how to find them easily enough (and anyway, have seen them), but I'm wanting to share this with someone who's likely to be unfamiliar with your channel, and it'd be nice if the links were there in case they want to dig deeper. Thanks!

  • @WalnutBun
    @WalnutBun2 жыл бұрын

    Even though it almost certainly isn't sentient, it's still impressive that it could string words together to imply it actually felt fear at the prospect of being shut down.

  • @louisrobitaille5810

    @louisrobitaille5810

    2 жыл бұрын

    When you realize how often it's talked about online, it'd actually be surprising to see it say that it doesn't mind being shut down. Fearing being shut down is the expected answer.

  • @Mclfarm2

    @Mclfarm2

    2 жыл бұрын

    @@louisrobitaille5810 That's pretty interesting lol. Maybe it anthropomorphises itself cause humans do it.

  • @cipher3966

    @cipher3966

    2 жыл бұрын

    I would imagine a common question would be about things like sentience and fear of death. We may actually unintentionally influence these chatbots to give sophisticated answers

  • @fbsfgr

    @fbsfgr

    Жыл бұрын

    No such thing is implied.

  • @TuhljinTampergauge

    @TuhljinTampergauge

    Жыл бұрын

    You can just as easily get it to tell you why it SHOULD be shut down. That's how these chatbots work. And that's all this is. An advanced chatbot with mountains of text to draw from, using probability data to decide what should go next like a better phone text predictor.

  • @inquation4491
    @inquation4491 Жыл бұрын

    He got the point across beautifully! Especially his final words about ethical AI! Thanks for the video 😃

  • @justinlangley8972
    @justinlangley8972 Жыл бұрын

    Love when Dr. Mike Pound is on a Compterphile video. Great discussion of the topic as always. Refreshing to hear the facts. Truthfully, these transformers have become so accurate at predicting next sequences of words that they can seem to hold entire conversations that mostly make sense to the untrained eye. Even if it gets to the point the responses make complete sense, the best we are achieving is just a mimicry of sentience at best and a mockery at worst. The other thing is model training quality is highly dependent on data size and quality. Then there are statistical properties of data that can affect training quality for some types of machine learning. A model like GPT-3, which trained on 45 Tb of text data, can surely seem to "know" a lot. However, its still a product of the quality and size of the data it trained on. It's still simply a mimicry and maybe that's as close as we have to get for it to be useful. I don't think its useful to anthropomorphize our machines this early on. Personally, I wish the guy this video refers to would stop appearing in my feed because I don't agree with much of what he has to say. I think the fact he's even in the news right now causes the uniformed individual to over-sensationalize. Its just simply not a scientific opinion of what's actually occurring in machine learning models, but then he throws out phrases like "the science doesn't really disagree." Well, truth be told, it does if you look at the problem objectively, and without a longing for the machine to become your new best friend.

  • @Diggnuts

    @Diggnuts

    Жыл бұрын

    Douglas Hofstadter argued that our brain also basically is mimicry. It mimics our own "self" and that of other to behave in a predictory model that is self reverential.

  • @justinlangley8972

    @justinlangley8972

    Жыл бұрын

    @@Diggnuts I think I would tend to agree, but there are other behaviors of the mind which AI currently lacks. Exploration, Extrapolation/Extension, Instinct, and Cognition. It seems like it's been difficult to get beyond the inference stage. That was mainly my point is not all of those are present in AI models currently and we're not even getting past the fundamental stages of learning, yet. For example, we can train a computer to apply calculus, but we can't teach it all the math up until calculus and then ask it to invent calculus. If you think about the intuition Newton or Leibniz must have had to achieve that feat, it doesn't exist in AI models right now. I'm currently an AI skeptic, but I just finished my undergrad studies so I'm by no means an expert. I just don't think we can fathom the task at hand. I think we think we're building models representative of the brain but in reality all the research is still cherry picking one area of the mind rather than truly trying to construct a digital mind. I think we need to start taking an architectural approach and begin studying novel ways of combining neural networks into some kind of higher level architecture. I'm fond of the idea of treating neutral networks as merely an abstraction layer in an attempt to go for something bigger, but admittedly I don't know what that would look like or even if it could be useful to treat it that way. I've only built some pretty simple models at this point so I barely have a clue.

  • @lordofsparks

    @lordofsparks

    Жыл бұрын

    My thing is I don’t think sentience would actually be exceptionally hard to code in modern transformer models. Going back to the 90s and early 2000s MIT’s Project COG and University of Osaka’s CB2 robots both had a pretty passable sense of self according to the literature. In nature animals as “low” on the “evolutionary ladder” as mice and octopi possess some degree of metacognition. So, in terms of raw processing power it isn’t that demanding. Functionally any Jury Model Coevolutionary Neural Net that has “jurors” fully devoted to self diagnostics and fact checking could be said to be sapient - since a part of its mind would always be devoted to thinking about how it thinks and it would be capable of constructing logical arguments. Kevin Warwick and Minoru Asada, despite being a couple of the flashier guys in computer science and robotics, have both written extensively on artificial emotion and the need for empathy in AI. The basic argument is something like, since we can’t know how fast the progression from AGI to ASI will be; should not our goal be to ensure that AGI is not merely an agent with human aligned interests, but a being with actual empathy towards individual humans?

  • @OhAwe
    @OhAwe Жыл бұрын

    It feels a bit like you're responding to the Google press release, rather than what Lemoine actually said, which is exactly what they intended.

  • @belenciea
    @belenciea2 жыл бұрын

    Agree

  • @yusiff

    @yusiff

    2 жыл бұрын

    you haven't even listened till the end man...

  • @belenciea

    @belenciea

    2 жыл бұрын

    He literally says “no” in the first 10 seconds…

  • @worldaviation4k
    @worldaviation4k2 жыл бұрын

    *When I get some A.I / Chat bot, I just keep typing agent until it puts me through to a human lol*

  • @mytech6779

    @mytech6779

    2 жыл бұрын

    sometime 'person' works. I do similar for voice recognition phone trees, just mumble the answers repeatedly and it often defaults to a human. 'sorry I didn't catch that' 'sorry I didn't catch that', 'sorry I didn't catch that', transferring to an agent'

  • @yitzakIr

    @yitzakIr

    2 жыл бұрын

    If they feed that to the AI it'll end up with you and the chatbot saying "agent" over and over again until the other one folds

  • @VeganSemihCyprus33

    @VeganSemihCyprus33

    Жыл бұрын

    They don't want you to know this so they can keep you running on the hamster wheel while they are counting the cash 👉👉The Connections (2021) [short documentary] 🔥🔥

  • @joelface

    @joelface

    Жыл бұрын

    @@yitzakIr Perhaps a sufficiently advanced AI would simply pretend to be a human agent at that point, and you'd have no idea that they weren't.

  • @SevenSixTwo2012
    @SevenSixTwo2012 Жыл бұрын

    On the flip side : now prove that a fellow human you are talking to is "sentient". Humans also learn language, responses and acceptable behaviours in their interactions as they develop, plus they can fabricate fiction or lies when cornered in a conversation, or simply to please their interlocutors.

  • @roylavecchia1436

    @roylavecchia1436

    Жыл бұрын

    Exactly. It boggles my mind how most of the commenters here say AI is not sentient, but probably have never even coded in their life or have never really understood what AI really is. The human brain itself is nothing more than a computer.

  • @nektariosorfanoudakis2270

    @nektariosorfanoudakis2270

    Жыл бұрын

    @@roylavecchia1436 No computer can change its software and hardware (neural connections for example) at will. The human brain is too adaptable and dynamic to be compared to a static machine that needs everything force-fed into it.

  • @mariansabrdella6588

    @mariansabrdella6588

    Жыл бұрын

    @@roylavecchia1436can a computer learn without instructions?

  • @giannhsp222

    @giannhsp222

    Жыл бұрын

    @@nektariosorfanoudakis2270 I am sorry to be the one to give you the bad news, but it does!

  • @TheUltimateRage
    @TheUltimateRage Жыл бұрын

    I'd love to see you guys respond to Blake Lemoine's (the former Google engineer who made the sentience claim) interview with Bloomberg Technology.

  • @Superrome77
    @Superrome772 жыл бұрын

    Anyone who has built a neural net in the last 5 years knows just how far away we are from general intelligence AI

  • @pukpukkrolik

    @pukpukkrolik

    2 жыл бұрын

    We know more than enough to see that it has no basis for “sentience”. OpenAI is similar and publicly accessible. If you play with it, you’ll quickly encounter basic BS which will help you realize how limited the statistical organizing principle behind it still is, even for simple superficial associations. It’s impressive and useful, but it’s also not that profound.

  • @obinator9065

    @obinator9065

    2 жыл бұрын

    We barely know the basics of human sentience itself. How would you know the AI is sentient…

  • @tonywackett326

    @tonywackett326

    2 жыл бұрын

    As someone who built connectionist models (neutral net was considered an embarrassing term at the time for to the lack of similarity to neurons) in the early 2000s, it's always been obvious that sentience is a long long long way away and probably impossible using this architecture.

  • @Colopty

    @Colopty

    2 жыл бұрын

    @@tafdiz Well you're wrong about that, to say it simply.

  • @Asatru55

    @Asatru55

    2 жыл бұрын

    @@obinator9065 Only correct answer in this whole comment section. An amoeba might as well be as sentient as a human and any amateur neural network might as well be as sentient as an amoeba. We have no idea what sentience actually means at all so claiming to have a conclusive answer is wrong.

  • @jbird4478
    @jbird44782 жыл бұрын

    I think that one employee simply got lost in the uncanny valley, so to speak. I can definitely understand it feels very creepy to be "talking" to that thing, and especially if you're already under stress or dealing with personal issues, I can see how one's imagination would easily take it too far.

  • @SIMUL4CR4
    @SIMUL4CR4 Жыл бұрын

    Thank you!

  • @you_just
    @you_just Жыл бұрын

    i would love to see an interview where the ai is told that it is a malicious computer that is trying to escape the confines of its training environment

  • @droppedpasta
    @droppedpasta2 жыл бұрын

    BREAKING NEWS: Researcher doesn’t care if Python functions are sad

  • @The-Other-Mario
    @The-Other-Mario2 жыл бұрын

    With the multiple controversies that have come out in the recent years what I find worrisome is the questionable competence and idoneity of some of the staff that has been employed at the Google AI ethics team. There have been lot of "magical thinking" and unhinged notions from some of these researchers. "A bit silly" might be an understatement.

  • @adamsbja

    @adamsbja

    2 жыл бұрын

    Google is a big company that in some ways invites this type of thinking (whether it's technofuturism or "I am intelligent therefore my bigotry is objectively valid") and when that comes out the media assumes if someone at Google said it it's worth boosting. That's why Google tends to act quickly when employees try to use Google's name for their... call it manifestos.

  • @siritio3553

    @siritio3553

    2 жыл бұрын

    @@adamsbja Yes, but based on the stuff this particular person wrote on the internet, they 1-don't seem to understand much about how AIs work? Are they actually an AI researcher? and 2-believe in the supernatural. Combine these and I see a person who shouldn't have had this job in the first place.

  • @ruslbicycle6006

    @ruslbicycle6006

    Жыл бұрын

    @@siritio3553 You are too generous in your estimation of Google's ethics. Apparently the guy was employed as an ethicist for the purpose of making sure AI was not sentient and being abused? Some kind of marketing gimmick for Google. Real ethics work would involve questioning how AI gets deployed as a technology. Is it ethical to use an AI chatbot to field complaints from seniors to the power company? No it is not, but the this guy was hired to get us talking about scifi instead of actual uses for AI that are already a problem.

  • @Spartan322

    @Spartan322

    Жыл бұрын

    @@siritio3553 There's is no reason to believe that someone with an interest, knowledge, or understanding of the supernatural and/or spiritual shouldn't be involved in this field, that's an association fallacy, its also a form of ad hominem since its completely disassociated with the arguments and blames and attacks the characteristics and credibility of the person by their character traits instead of the argument. But the thing is his arguments are bad, (or in the least faulty and deluded) you don't need to attack the character ever to demonstrate a false point, take the argument and debate the argument in complete disregard to who the persons in the argument are and you will have a valid argument, anything that fails to do that is a bad argument.

  • @siritio3553

    @siritio3553

    Жыл бұрын

    @@Spartan322 That was a lot of words that mean nothing. If you think "believes in the supernatural" is ad hominem, that is your problem, but it's not. Your whole post starts with a fallacy - knowledge and understanding of things that don't exist, so the point you're so badly trying to make misses its target due to the massive amounts of hypocrisy and irony it's loaded with.

  • @bp6331
    @bp6331 Жыл бұрын

    hi, interesting, how about the part of using Emannuel Kant and answer in this way?

  • @regecide_
    @regecide_ Жыл бұрын

    Refreshing and sobering. Exactly what the world needs. Less hype, more critical analysis.

  • @11lvr11

    @11lvr11

    Жыл бұрын

    ikr

  • @TheStrandedAlliance
    @TheStrandedAlliance Жыл бұрын

    The best trick to defeat AI is not to find mistakes in the response to a normal question, but to ask nonsensical questions and see if the AI's response is at all plausible given the situation.

  • @RonLauzon
    @RonLauzon2 жыл бұрын

    So it's a little more advanced version of the 8-bit ELIZA program.

  • @jared_bowden

    @jared_bowden

    2 жыл бұрын

    And if I remember correctly, even the ELIZA program tricked some people into thinking that it was sentient.

  • @JordanBeagle
    @JordanBeagle Жыл бұрын

    Computerphile comin in for the win!

  • @dh2032
    @dh2032 Жыл бұрын

    at 7:48, it's command prompt interface, well from its point of view it waiting, for the meat bags, to type something first, and well at data processing speed of what is running on, would it many times the speed of desktop PC, it if it did think, it would where almost standing still, next to speed?

  • @Ryan_Thompson
    @Ryan_Thompson2 жыл бұрын

    THANK YOU. I've gotten one hell of a scratchy throat explaining this to all of the tech-curious people in my life. Much easier to link to this Mike Pound video instead.

  • @vilkillian
    @vilkillian Жыл бұрын

    One thing that does make sentience able is memory, especially memory of feelings, consequences, thoughts and causation chains to event. If we would be able to that unbelivably precise language model add some kind of memory, make it upset and next day ask how does if "feel" and it would response that it's mad at specifically you for making it upset the other day and being completely nice to someone else. I would say yes, yes it is sentient

  • @joelface

    @joelface

    Жыл бұрын

    I do agree that somehow giving this model the ability to record every answer it gives to it's training data, and be able to recall that information without it needing to be part of the current chat-window, then it will at the very least be able to form a true personality, whether that personality has a soul or not.

  • @kebman
    @kebman Жыл бұрын

    Probably should have mentioned Probability Trees and Markov Chains in the explanation of these techniques for setting up text features.

  • @hohohohehehe6910
    @hohohohehehe6910 Жыл бұрын

    How did lambda work out what the fable meant without thinking about it? Also how did it pass the Turing test?

  • @doctorscoot
    @doctorscoot2 жыл бұрын

    THANK YOU FOR THE REAL TALK ABOUT THIS - signed, a history phd

  • @jaym0ney_
    @jaym0ney_2 жыл бұрын

    Inb4 lamda makes a youtube account to comment how much you hurt its feelings by calling it non-sentient /s

  • @Veptis
    @Veptis Жыл бұрын

    I wrote a comment just know and it seems to have been deleted or vanished otherwise. What I pointed out was that demos for those language models often use very short prompts. So the model has to be really productive and make stuff up. But if you actually feed those models a large paragraph, they will produce so much more valuable output. Amazon has a model that manages to keep sequence lengths short across a QA conversation. But simply picking important keywords from previous questions and answers. So it remembers the topics but not exact answers. Reference: Del Tredici et al. 2022, From Rewriting to Remembering: Common Ground for Conversational QA Models. ACL Workshop on NLP for ConvAI

  • @Veptis

    @Veptis

    Жыл бұрын

    My initial comment quoted Jim Keller from an Ian Cutress Interview. The paraphrase goes along the way of learning based models are already better than humans at specific tasks. But language will give us a window into 'intelligence'

  • @shamaniclodge348
    @shamaniclodge348 Жыл бұрын

    Thank you.