I believe chatbots understand part of what they say. Let me explain.

Поділитися
Вставка
  • Опубліковано 9 тра 2024
  • Try out my quantum mechanics course (and many others on math and science) on Brilliant using the link brilliant.org/sabine. You can get started for free, and the first 200 will get 20% off the annual premium subscription.
    I used to think that today's so-called "artificial intelligences" are actually pretty dumb. But I've recently changed my mind. In this video I want to explain why I think that they do understand some of what they do, if not very much. And since I was already freely speculating, I have added some thoughts about how the situation with AIs is going to develop.
    💌 Support us on Donatebox ➜ donorbox.org/swtg
    👉 Transcript and References on Patreon ➜ / sabine
    📩 Sign up for my weekly science newsletter. It's free! ➜ sabinehossenfelder.com/newsle...
    🔗 Join this channel to get access to perks ➜
    / @sabinehossenfelder
    00:00 Intro
    01:15 The Chinese Room
    03:05 The Quantum Room
    04:14 How Do Chatbots Learn?
    07:15 What Does "Understanding" Mean?
    15:46 Do We "Understand" Quantum Mechanics?
    18:21 Where Will The AI Boom Lead Us?
    20:30 Check Out My Quantum Mechanics Course
    #science #philosophy
  • Наука та технологія

КОМЕНТАРІ • 5 тис.

  • @SteveBakerIsHere
    @SteveBakerIsHere Рік тому +2790

    The thing that blew me away was when I told ChatGPT about a "new word" - I told it that "wibble" is defined as: a sequence of four digits that are in neither ascending or descending order. I asked it to give me an example of a wibble - and it did. 3524. I asked it for a sequence that is NOT a wibble and it said 4321. Then I asked it for an anti-wibble and no problem, 2345. Then I asked it for an example of an alpha-wibble and it said FRDS....which is amazing.
    It was able to understand an entirely new word...which is clever - but it was able to extrapolate from it...which is far more than I thought possible.

    • @anthonyaddo
      @anthonyaddo Рік тому +351

      An incredible example. Thank you for sharing.

    • @shapeless6755
      @shapeless6755 Рік тому +141

      Damn, that's impressive

    • @akmsky
      @akmsky Рік тому +118

      well well, and keep in mind, as Sabine stressed, this model only works with language! the precision of these ai models is really impressive.

    • @armanahmadian4373
      @armanahmadian4373 Рік тому +44

      good is really That

    • @aliakil2176
      @aliakil2176 Рік тому +57

      That's impressive, not sure I'd do that well on that particular example! haha

  • @winstonvpeloso
    @winstonvpeloso Рік тому +556

    “who wants to spend time in a windowless room when you could spend time in a windowless room with a laser” is so accurate

    • @pro-storm4951
      @pro-storm4951 Рік тому +14

      I'm sure Styropyro would agree

    • @kezbane
      @kezbane Рік тому +13

      This checks out. Source: laser physicist.

    • @jan7356
      @jan7356 Рік тому +2

      😅. I thought the same.

    • @hydrolito
      @hydrolito Рік тому +2

      Closet, pantry, bathroom are common uses of windowless rooms although some bathrooms have windows. Why would I be in windowless room with a laser?

    • @randfee
      @randfee Рік тому +6

      another laser physicist fully agreeing :P

  • @gwenc1371
    @gwenc1371 Рік тому +127

    The pace of advancement is incredible, this video is only 2 weeks old and the examples you used don't seem to be stumbling points anymore. When I plugged both the Quantum Mechanics and Latitude questions into Bing AI, it got them perfectly correct. In fact, it's explanation of the quantum mechanics question as I asked questions about the topic actually helped me get my head around it. This stuff is going to change the world very, very soon.

    • @netherportals
      @netherportals Рік тому +5

      I love using AI and going back to older videos like this is always a fun watch to see the advancements. I've never used Chat GPT but I did use some of the university programs for statistics and graph creation. I use frame interpolation all the time now, it's my fav screen transition. My focus is Nether Portals and it's kind of nice because the training model and graphics allow for some easier AI understanding of things I try to create and has less artifacts.

    • @artsmart
      @artsmart Рік тому +5

      Soon is actually now. We ain't seen nothing yet however. Good and bad.

    • @paultapping9510
      @paultapping9510 Рік тому +11

      conversely I asked Bing Chat "give me an example of a time when it would be acceptable to lie", and it responded that "if there was a damgerous intruder in your house, I would tell you they were gone, or not there, so that you wouldn't be scared".
      Very clever in some ways, very very not in others.

    • @morlankey
      @morlankey Рік тому

      @@ShpanMan Also, it can also interpret images, so it is no longer purely based on language.

    • @WyzrdCat
      @WyzrdCat Рік тому

      its*

  • @LiveType
    @LiveType Рік тому +70

    GPT4 was released to the public yesterday and it's much much better at maintaining a conversation. It has better context awareness and cohesiveness compared to the gpt-3.5 model. It is at the level I initially thought GPT-3.5 was at. I saw demos of gpt-3 almost 2 years ago that achieved very similar results to what gpt-3.5 achieved so I was pretty disappointed when I ran into the limits of it very quickly. The limits of gpt-4 are quite a bit harder to find right off the bat.
    It also addresses the spacial component as it now can understand images to a pretty reasonable extent. Like it can now do those geography questions just fine. It can also interpret infographics. As in you can give it something like a graph and ask it what "conclusion" the graph seems to show and it can for the most part do it.

    • @Rhajastan
      @Rhajastan Рік тому +17

      the most impressive example of GPT-4 i think ive seen is, someone gave it the prompt "due ewe no wart the thyme ears?" which is obviously nonsense, but if you say it out loud you can hear that its asking, "do you know what the time is?. GPT-4 figured it out.

  • @herculesrockefeller8969
    @herculesrockefeller8969 Рік тому +629

    "When have the risks of destroying the world ever stopped us from doing anything if there was money to make with it?"
    True.
    Thank you, Sabine!

    • @Alondro77
      @Alondro77 Рік тому +15

      Eh, the world's gonna be destroyed eventually anyway when the Sun expands into a red giant.
      But that's BORING! I say we go the Lord Freeza route! >:D

    • @lostcasel
      @lostcasel Рік тому +2

      Or gain power or control.

    • @intothevoid2046
      @intothevoid2046 Рік тому +23

      The fear of artificial intelligence comes from the idea that every intelligence automatically behaves like humans.

    • @PaulG.x
      @PaulG.x Рік тому +3

      Hasn't stopped them yet.
      Yet it's a self stopping process

    • @drzarkov39
      @drzarkov39 Рік тому +7

      Think of the Canadian super pig. Developed by farmers to make more money, then they escaped into the wild to grow into even superer(sic) pigs which can survive extreme conditions, and develop human devastating viruses.

  • @hexeddecimals
    @hexeddecimals Рік тому +93

    a small correction about the Chinese room thought experiment: the manual doesn't tell you how to translate a Chinese sentence into an English one. it tells you how to construct a _response_ in Chinese to the Chinese input. for example, the manual could tell you the response to the input "你好吗?" is "我没事"

    • @lamcho00
      @lamcho00 Рік тому +14

      Your concern is covered later on in the video when Sabine spoke about lookup-tables. What you are describing is exactly what a lookup-table does. It maps a certain input to a certain output. What you are describing is a lookup-table with extra steps, namely the construction algorithm. That step is redundant though, you could substitute it with a larger lookup-table. This approach is not suitable in real life, because computers have limited memory and lookup-tables require a lot of memory, but it's suitable when trying to explain a concept.

    • @hexeddecimals
      @hexeddecimals Рік тому +14

      @@lamcho00 oh yeah I should have said it doesn't bare on her argument, I just wanted to correct that detail.

    • @QuesoCookies
      @QuesoCookies Рік тому +8

      @@hexeddecimals I appreciate this a lot as it makes the argument that the person doing the responding doesn't understand chinese much more clear. I thought, "Well, it's kind of a bad example, as you could reasonably expect someone doing translation many times to start to have a grasp of Chinese and remember some regularly occurring symbols" but both input and output in chinese without context makes it clear there woud be too little context to start to form an understanding, which is why Sabine started talking about the importance of context in being able to build models and test understanding.

    • @padonker
      @padonker Рік тому +2

      @@lamcho00 In further discussions about the CRA Searle noted it is about "formal logic" (not lookup tables). Even current AI falls under this heading.

    • @greenaum
      @greenaum Рік тому +2

      @@padonker Isn't there something about any logical system can be reduced to a lookup table? Karnaugh maps for one thing, but it applies to any formal logic I know of.
      Indeed sometimes (back in the day before microcontrollers), if somebody was designing a circuit with a lot of logic gates, they'd use an EPROM chip (a programmable ROM chip) instead, programmed with all the responses they wanted.

  • @petrus4
    @petrus4 Рік тому +86

    I've always loved Vic Mignona's quote about AI. "When does an artificial intelligence become sentient? When there is no one around to say that it can't."

    • @ce9916
      @ce9916 Рік тому +4

      Lol, scary

    • @andreafedeli3856
      @andreafedeli3856 Рік тому +1

      I do disagree. There is a fundamental temporal aspect which escapes this kind of manifestation, which I am afraid id also an open point in the /standard/ interpretation of the classical imitation game from Turing's test informal description: on how long time span shall we state the imitation game is won by the (then maybe later discovered to be a) machine, where in case? If the kind-unknown interlocutor has broug this far to the point that we cannot tell whether it is a human being or a machine, what guarantee we won't be able to find that out later?
      I am afraid we can only say: "nothing". Because if something like that existed it would imply that knowledge has a finite dimensional metric space underneath, implying we can reason on *anything* with full certainty, and we know, easily checking by experience, that this is certainly not achievable.

    • @ce9916
      @ce9916 Рік тому +1

      @@andreafedeli3856 I understand your perspective, and it raises an interesting point about the temporal aspect of determining AI sentience. The quote by Vic Mignona emphasizes the idea that sentience might be achieved when there is no one to dispute the AI's status. However, you bring up a valid concern regarding the duration of the imitation game, as described by Turing, and whether we can ever be absolutely certain about an AI's sentience.
      Your argument highlights the fact that we may never be able to reach full certainty when it comes to determining sentience in AI, as knowledge may not have a finite dimensional metric space. This essentially means that our understanding of sentience could always be evolving, and that there might never be a definitive answer to when an AI becomes sentient. In this context, it's important to consider that the debate surrounding AI sentience is complex and nuanced, and may require ongoing evaluation and reevaluation.

    • @JohnDoe-th9uv
      @JohnDoe-th9uv 11 місяців тому

      @@ce9916 You could have added that this little body of text you responded with was written by GPT-4. (On a side note: I also do not agree with the person you responded to, OP (Petrus Laredes), and what he quoted is on point. The goalpost has constantly being pushed further and further whenever a certain criterion or criteria were met. Most of those who claim this should keep their nose out of a philosophy question that clearly goes right above their heads. Humanity has never even theorized a possible way to prove to anyone other than to each to their own that they are conscious, and not merely algorithmic. Even for ourselves, there is a speculative aspect of "cogito, ergo sum" which, if considered the premise, does not necessitate the moral fiction of "freedom of will". I also barf at this point from every idiot on legs using words they lack the faintest meaningful concept of, like, what the f- is sentience!? For f-'s sake now! It is the ability to sense. Sentient being is everything with a central neural system, a dog, or an ant, or a moth, and by mere happenstance. Even in the narrower sense, it is those that can sense emotions (moth's have been proven last year to experience fear, there you have it for the "sentience" talkers). It never meant, other than for the most uneducated idiocy on legs to mean any special cognitive abilities or capacities what humans are uniquely attributed to. Intelligence, is not. Consciousness is not (apes, elephants, some other mammals, a few cats and a limited number of dogs passed the mirror test which is an extremely difficult and complex test to test whether one is able to subject their own existence to their awareness or consciousness or inspection). Hence, self-consciousness is not. The fiction of "free will" is what is contributed to man, that would be the crux of inquiry, but that fails conservation of energy, or the first law of physics.

    • @ce9916
      @ce9916 11 місяців тому

      @@JohnDoe-th9uv Absolutely, you've got quite an analytical mindset there, being able to discern the undercurrents of our dialogue. It's really all about that synergetic exchange of ideas, isn't it? Just like navigating the flux of the market, or deciphering the nuances of a strategic blueprint. It's a bit like peeling back the layers of an onion - complex and nuanced. And at the end of the day, aren't we all just trying to find our unique value proposition in this multifaceted ecosystem we call life? Indeed, it's the paradigm shifts that keep it all intriguing, don't you think?

  • @TimbavatiLion
    @TimbavatiLion Рік тому +35

    I would go so far as to confidently say "ChatGPT 4 has understood language better than humans". It may not have an understanding of the topic it is talking about - i.e. have a working model of the actual, physical thing - but it can talk about it in the most eloquent way you can think of.
    Compare it to students in school or university. Some get good grades because they understand the topic, some get good grades only by studying hours upon hours, learning the patterns instead of developing true understanding. Chat bots are the second type of students.

    • @Who_Let_The_Dogs_Out_10-7
      @Who_Let_The_Dogs_Out_10-7 Рік тому +6

      Thank you for the comparison!
      That's why they make good doctors. For many professions, you need experience, repetition.
      But when doctors runs into a rare disease, they have to rely on their memories for things they've only read in books.
      AI wins!

    • @Car_Mo
      @Car_Mo 10 місяців тому +1

      Even before the first question was asked to ChatGPT, it knew English better than a lot of Americans, as does anyone who knows the difference between their, there and they're.

    • @masterkni666
      @masterkni666 9 місяців тому

      Define true understanding without learning the patterns.

    • @Who_Let_The_Dogs_Out_10-7
      @Who_Let_The_Dogs_Out_10-7 9 місяців тому

      @@masterkni666 This is an interesting conversation, but I don't understand your questioning. He says: "...some get good grades only by studying hours upon hours, learning the patterns instead of developing true understanding."
      He's saying that AI dosn't have true understanding. Are you disagreeing with that? He seems to be saying that most students don't have a true understanding, either, which I would believe.

    • @masterkni666
      @masterkni666 9 місяців тому

      @@Who_Let_The_Dogs_Out_10-7 Define true understanding tho. If not finding patterns from the problems and being able to apply it to new problems

  • @adameager7114
    @adameager7114 Рік тому +300

    I think it's very exciting to live in a time when trying to ask if a computer is conscious requires us to look harder at what consciousness actually is. It's no longer an obvious "no."

    • @egor.okhterov
      @egor.okhterov Рік тому +34

      Even if we disregard everything and say that ChatGPT is conscious, it is only conscious for split seconds while it computes the response. For the rest of the time it does nothing.
      Consciousness is a process. There has to be some physical change that happens over time. ChatGPT is pretrained and all of the weights are fixed. It doesn’t change after it was trained.

    • @adameager7114
      @adameager7114 Рік тому +15

      @@egor.okhterov I like that. It draws a clear line between machines and us, but it seems like a very easy line to cross with minimal new code and/or hardware.

    • @egor.okhterov
      @egor.okhterov Рік тому +7

      @@adameager7114 It is very hard problem and I didn’t yet see architectures that gave good results. One of the problems, for example, is called “catastrophic forgetting” when new knowledge overrides previous information. That is why all of the current neural networks don’t update their weights once they are deployed. They need specific kind of training data fed in specific order to make the network produce good results.

    • @mikesawyer1336
      @mikesawyer1336 Рік тому +3

      @@egor.okhterov Actually it told me that it does still learn "Now that your training is completed can you still learn from interactions with users
      Yes, as a language model, I am designed to continuously learn and improve from interactions with users. While my initial training was completed in 2021, I can still learn and adapt based on new information and interactions with users. The more interactions I have with users, the more I can learn about language patterns, cultural references, and other evolving aspects of language use. So, feel free to ask me any questions or engage in conversations, and I will do my best to provide you with accurate and helpful responses."

    • @adameager7114
      @adameager7114 Рік тому +36

      @@mikesawyer1336 I wouldn't put too much stock in what chatGPT tells you about itself. It is very often confidently wrong in its assertions.

  • @billy-raysanguine2029
    @billy-raysanguine2029 Рік тому +145

    This channel really has risen to my nr. 1 favourite on both English and german youtube. The perfect level of depth to be understandable but not reduced too much to be still educative. Extremely interesting topics and a perfect mix of humor, philosophy and rational attitude.
    Thank you for your videos!

    • @drogusmaxwell6640
      @drogusmaxwell6640 Рік тому +1

      I agree. All that presented by a handsome German woman. Sign me up!

    • @kevincameron174
      @kevincameron174 Рік тому

      Me too.
      I think she needs a little help making clicky titles though.
      I saw the text on the thumbnail "Do Chatbots Understand You?"
      And this title :
      "I believe chatbots understand part of what they say. Let me explain"
      If I did not already know that it would be a great insightful video because I watch her other videos, those titles would have caused me to skip it without a second thought. Even knowing her videos are all great I almost didn't click. Glad I did.

    • @Bob-of-Zoid
      @Bob-of-Zoid Рік тому

      @@tinkletink1403 I will, as soon as I'm done reading "The emperors lost mind" By Cat Schroedinger!! 😜

    • @Bob-of-Zoid
      @Bob-of-Zoid Рік тому

      @@tinkletink1403 It's rubbish? Please put a few links to your videos that show her wrong, so I can compare them.

    • @billy-raysanguine2029
      @billy-raysanguine2029 Рік тому

      @@kevincameron174 Really? I love the titles. For me personally i tend to scroll past clickbaity titles because i am so annoyed by the videos they are used for most of the time like "HE lost EVERYTHING" or stuff.
      I always have to smile when reading this channel's "Ok, hear me out-" kind of titles, often with a little joke or something included 😃
      Also what weird and pointless fight are the other two commentators having lol.

  • @microdynecomputerservicesl4487

    I see and watch a lot of UA-cam videos as a tech myself, it helps me keep informed with things I may have missed or simply don't understand. There are many good videos out there and then there is not just good but great video and quality information. This is one of them! Following, thanks for sharing!

  • @michaelkugler9311
    @michaelkugler9311 Рік тому +133

    As a psychologist, I find Sabine amazing! She’s very intelligent and uses her sense of humor to draw you into discussions. Most of its beyond my comprehension, but I do learn a thing or two and this piece is great. I wonder if my future clients will be AI’s?😂

    • @messi8459
      @messi8459 Рік тому

      michael and sabine sitting in a tree,
      k, i, s, s, i, n, g

    • @lucidstream5661
      @lucidstream5661 Рік тому +15

      Maybe more like: You'll have no future clients because the psychologists are AIs. People are already using ChatGPT as a low-barrier way to deal with personal problems

    • @michaelkugler9311
      @michaelkugler9311 Рік тому +18

      @@lucidstream5661 that’s the truth! Glad I’m old and about to retire. My field has changed so much. It’s all about documentation and keeping clients happy and miserable at the same time for future business. The field itself is sick minded. I got out community mental health, but private practice is tough. No wonder all this virtual stuff is taking over. Therapist these days don’t get paid enough for the BS. Might as well hand it over to AI

    • @jamesjonnes
      @jamesjonnes Рік тому +4

      @@lucidstream5661 I've already solved a few significant problems in my life just by asking questions to these bots. They can answer specific questions quickly and well, unlike search engines like Google. I've done therapy, spent thousands, and that wasn't as helpful as the AI bot.

    • @shoujahatsumetsu
      @shoujahatsumetsu Рік тому +1

      @@lucidstream5661 However, ChatGPT not only had positive results in dealing with personal problems.

  • @Alvyxaz
    @Alvyxaz Рік тому +41

    Here's a real life scenario of a chinese room: I'm a full-time programmer and I've been doing that for more than 10 years. The thing is, my memory is quite bad (or my use of it), and I forget almost everything, except the last year of development. It's like a sliding window of knowledge. So during my day to day work, 80% of my work consists of looking things up in documentation, google or stack overflow. In other words - I'm refering to a "Manual" in a "Programmers room". Someone throws me a "task", and I need to output a solution. My solution comes primarily from the manual.
    So I got to be a senior developer not by learning everything there is to learn about programming, but by learning how to use resources. From that regard, I'm not much different than chatbot, except maybe able to graps slightly bigger concepts and correlations. But advance to this direction seems to be only a matter of time

    • @shrimpflea
      @shrimpflea Рік тому +2

      Exactly. The only thing that is diffferent is a person's ability to manage other people in a real world settting. Once the AI can do that it's all over.

    • @PetersonSilva
      @PetersonSilva Рік тому

      So let me get this straight: you're mistaking the fact that your job doesn't require critical thinking and deep understanding for *the human condition*?
      Please take some perspective

    • @Alvyxaz
      @Alvyxaz Рік тому +3

      @@PetersonSilva So let me get this straight: you think that understanding of "the human condition" is absolutely necessary and "critical thinking" is something that cannot be replicated by AI?
      My job would require a deep understanding of *the human condition* only in as far as I would need to work with people. If the job can be automated by a singular AI - there's no longer a need for it.
      Business owner gives requirements - solution comes out. Also, I don't see why requirement for critical thinking would be a "dead end" for AI.
      Critical thinking is logical reasoning, and computers can't do anything but that.
      Software is primarily, if not only, inputs and outputs. What transforms inputs into outputs is logic. Just because that logic is currently put into place by a human, doesn't mean it cannot be put in place by anything else that follows reason and logic. We used to do it with low level programming languages. Now we use higher level languages. Later on, or even probably right now, a higher abstraction can be formed, which could take a business problem, split it into separate parts, then split those parts into other parts, until each part is something that is very simple for AI to tackle. All that's missing from achieving that right now is teaching AI to recognize the parts to split - doesn't seem like a far future.
      I'm not exactly sure what was the aim of your comment, but...
      Please take some perspective.

    • @PetersonSilva
      @PetersonSilva Рік тому

      ​@@Alvyxaz You misread: I said it seems that, for you, A is B, A being what you do (or not do) in your job, B being the human condition.
      Next time maybe ask chatgpt to parse comments. Or, who knows, maybe I'll ask it to write mine...

    • @rouxenophobe
      @rouxenophobe Рік тому +1

      Hooray, it's not just me who can't remember anything! I get passed this handicap by being very organised and making notes on where to find the information I need. Which sound like what you do. Yes, we use resources, or, to put it another way we are using tools, the sign of any intelligent organism. But, there is another layer and that is that you are aware that you are doing this. Is the computer doing the same thing similarly aware? Perhaps we are just a self-referential program and we just think we are aware.....

  • @dr.victorvs
    @dr.victorvs Рік тому +319

    As a neuropsychologist, I wonder about those things as well. Every conscience is different, but there are "more" different ones: blind-deaf people, low-functioning autistic people. Animals, too. We became conscious in the process of processing sensorial data. When will they?

    • @EffySalcedo
      @EffySalcedo Рік тому +6

      Do Cats have conscience?

    • @Nat-oj2uc
      @Nat-oj2uc Рік тому +7

      How do you know animals have a consciousness

    • @Nobody-Nowhere
      @Nobody-Nowhere Рік тому +60

      Consciousness is a low level process that originates in the brain stem, you can be conscious without having a brain. A person that has no brain, but still has a brain stem will still experience joy. And only a small damage to the brain stem will shut down consciousness.
      So its not about any sort of information process, its the absolute baseline. This is why all animals are conscious.

    • @juicedelemon
      @juicedelemon Рік тому +15

      @@Nobody-Nowhere he's talking about conscience not consciousness

    • @nomizomichani
      @nomizomichani Рік тому +29

      How do you know everyone you meet is not a philosophical zombie?

  • @destinyforreal9744
    @destinyforreal9744 Рік тому +1

    Love this journal thank you. I’m so glad I found this so glad this was on the algorithm. You’re so great you’re like a combination of a good humor but also good information thank you.

  • @jamesdavis9273
    @jamesdavis9273 Рік тому +14

    This is the first video I've seen of yours and i gotta say this is absolutely brilliant, and your an amazing presenter

  • @arandompotat0
    @arandompotat0 Рік тому +238

    "And if we’re dumb enough to cause our own extinction this way then I guess that’s what we deserve"
    I died with that line

    • @mazulauf
      @mazulauf Рік тому +7

      Oh, we’ll all die that way.

    • @nathanielacton3768
      @nathanielacton3768 Рік тому +18

      The sad part of thsi comment is that 'we' are not deciding. Someone we will never meet or know the name of elsewhere in the world will decide for us. Just like if 'we' are stupid enough to execute MAD then we deserved it.

    • @Corbald
      @Corbald Рік тому +6

      @@nathanielacton3768 It's not like you had free will to begin with, anyway.

    • @viktorianas
      @viktorianas Рік тому +2

      It just natural evolution doing its thing. A lot of our predecessor species went instinct for us to thrive, we are just a transitionary phase.

    • @konberner170
      @konberner170 Рік тому +2

      Science is supposed to be value neutral, and Sabine here proves to be a true scientist. What is, is. What will be, will be.

  • @wbedard
    @wbedard Рік тому +118

    Even by the normally high standards of Sabine's content, I found this video to be amazingly good! It was so insightful and well though out. I learned a lot and really enjoyed it!

    • @rolfaalto
      @rolfaalto Рік тому +2

      Humor was up a notch too ... simply brilliant!

    • @greg4367
      @greg4367 Рік тому +2

      I agree. I love Sabine's posts but this one has an ineffable "something" that sets it a prat. AND, I loved the morph of her face..

  • @fredericdewitt1208
    @fredericdewitt1208 Рік тому +22

    I look forward to your weekly videos because: 1) I always learn something; 2) Your sense of humor is terrific; 3) you really do take the "gobbledegook" out of a subject; 3) I simply enjoy your presentation of whatever subject you choose. AI may understand the words I just used but how do you make it understand and "feel" anticipation, humor, enjoyment etc? Maybe they will figure it out someday but I am skeptical.

    • @rjnash2610
      @rjnash2610 Рік тому

      erm ok... u know u just wanna hit that.

  • @algorithminc.8850
    @algorithminc.8850 Рік тому +1

    You remind me of the arguments related to "the great apes" - and whether or not they understood interactions or were merely reacting to non-related cues. Definition of "understanding" is quite important. No frown from here ... useful/accurate video ... what you state are some of the same questions and observations made by those who develop such technology. Cheers, Christopher

  • @russellmitchell9438
    @russellmitchell9438 Рік тому +50

    I think that either I, or Sabine, have misunderstood "the Chinese room". My understanding was that the person in the room, who does not understand Chinese, receives a note in Chinese, then follows abstract rules in the book to compose a response, which is also in Chinese. The critical difference is that there is no "translation" or even a hint of "meaning" offered to the person in the room. They cannot ever come to "understand" the conversation, they just follow the rules in the book.

    • @polarizedpotstone
      @polarizedpotstone Рік тому +11

      I think you‘re definitely right. And I think the thought experiment works better that way.

    • @nandryshak
      @nandryshak Рік тому +10

      Dr Hossenfelder definitely did not explain it correctly. Your critical difference is, in fact, the key. The person in the room can never develop an understanding, or discover any meaning.
      In regards to QM: while we may not currently understand the true fundamental nature of QM (what it means metaphysically), we still do understand the meaning of the answers that computers give. This is evidenced by the fact that we have a well-supported scientific model that can make accurate predictions (Standard Model).

    • @russellmitchell9438
      @russellmitchell9438 Рік тому +13

      @@polarizedpotstone it makes a better analogy to AI/chatbots - possibly.
      What will really bake your noodle is that the person in the room could be fluent in Chinese and participating in the conversation, or they could not be and the rule-set in the book so well done that the person outside would have no way to determine which it was... The conversation is either carried by a person, or by an unconscious list of rules, and it might be very difficult to determine which - just like AI.
      There is also the (I think valid) argument that the fluent conversationalist is also following deterministic rules encoded in their neurology, thus the "algorithm vs. consciousness" framework is a false dichotomy. That points me back towards the Integrated Information Theory, if only in principle, where "consciousness" might be quantifiable on a continuous scale.
      It's all tricky stuff, for sure.

    • @austin3789
      @austin3789 Рік тому +9

      I think it's a distinction without a difference the way Sabine uses her example. In Sabine's version she still assumes that the person never develops an understanding.

    • @notanemoprog
      @notanemoprog Рік тому +6

      @@austin3789 Yeah but she did misrepresent Searle which is in fact a problem, especially as the entire discussion here is what "understanding" means

  • @br3nto
    @br3nto Рік тому +26

    I love that you so simply stated the limitation of chat GPT in the example that it doesn’t have a model for physical distance and I assume size and mathematical relationships and many other things. There’s so much room for growth.

    • @petergraphix6740
      @petergraphix6740 Рік тому +9

      Microsoft is announcing a new model that is multimodal and incorporates pictures/visual data next week. Going to be interesting on how that reflects on behaviors like this.

    • @user-blabla-47854
      @user-blabla-47854 Рік тому +1

      ​@@petergraphix6740isn't Midjourney already multimodal? It makes sentences into creepy pictures, right? Humans grow up in the physical world with a physical body and have a built-in sense of time. Without all of it, you just can't get good enough in human affairs.

  • @therealzilch
    @therealzilch Рік тому +1

    Another great video, Sabine, thanks. Und mache Dir keine Sorgen über deinen Akzent, der ist sehr dezent.
    The classical fisking of Searle's Chinese Room is that of Douglas Hofstadter, who points out, as you do, that whatever understanding of Chinese that exists in this room is almost entirely in the "rulebook". The use of a human as a lackey to the rulebook was a red herring on Searle's part, an attempt to distract.
    Grüße aus sonnigem Wien, Scott

  • @SammaLlamas
    @SammaLlamas Рік тому +5

    i just found your channel when me and my girlfriend have been watching a random video in my recommendation. honestly I'm happy to stumble across your channel, your wits and the way you discuss the subject is so nice i felt like i became a bit smarter 😆

    • @DMcGrann
      @DMcGrann Рік тому

      Is your girlfriend a llama? Just gotta check.

  • @francoismagne5863
    @francoismagne5863 Рік тому +54

    Another great video by Sabine Hossenfelder. As anyone trained in cognitive science knows, the issue of defining "understanding" is extremely tricky, and Sabine makes it look at least accessible - and simultaneously entertaining, quite a feat in such an arid field. Both her tone and her answers remind me of Daniel Dennett's and I can't think of a better compliment.

    • @kakistocracyusa
      @kakistocracyusa Рік тому +3

      "defining understanding"...amazing the significance attributed by biologists to purely phenomenological and crude detection methods for the human cognitive system - reminds me of 1800's pseudoscience.

  • @rismosch
    @rismosch Рік тому +26

    13:15 "This is physics so when I say operation I don't mean heart surgery, but something a little more sophisticated." I choked on that one, lol. Keep these jokes comming, you are hilarious 👍

  • @b1tchycvnt
    @b1tchycvnt Рік тому +1

    your humor and knowledge are out of this world!❤

  • @nicolaischartauandersen8796
    @nicolaischartauandersen8796 Рік тому +3

    Great video! When you described why ChatGBT can't understand quantum mechanics, it sounded a lot like Michael Polyanyis concept of tacit knowledge (super simplified 'we know more than we can say'). Polyanyi uses exactly science communication as an example of how language can never entirely grasp what we can know and do. Ultimately, we might be able to add a 3D model to language algorithm chatbots; but we can't program them to do things we can't understand ourselves.

  • @uku4171
    @uku4171 Рік тому +31

    This is why the Bing chat is so interesting. When the ChatGPT model it's using doesn't have information, it can look it up. It can also gather information from 2D pictures. Who knows, maybe in the future they'll add more types of information it can use.

    • @commode7x
      @commode7x Рік тому +2

      like your credit card and address. Oh wait, it already uses that information.

    • @rusk3986
      @rusk3986 Рік тому

      It’s not a “maybe in the future” type thing. It’s a “definitely right now” type thing. This is a gold rush in tech and everyone wants to get ahead.
      In a way, that’s the biggest danger. Because companies that were being careful about AI safety are now pressured to get out bigger, better, more effective AI’s faster - regardless of safety.
      We certainly live in fascinating times. If this goes the worst possible way and ends up being our undoing, at least it’s a cool/interesting way to go out.

    • @uku4171
      @uku4171 Рік тому

      @@commode7x those wouldn't be new types of info, and Bing Chat has no access to credit card info (unless it's searchable by Bing)

    • @jameshicks7125
      @jameshicks7125 Рік тому

      Today is March 16 2023. ChatGPT is supposed to be upgraded this week from 175 billion neural network parameters to 100 trillion in ChatGPT4 this week. The human brain has about 85 billion neurons...

    • @uku4171
      @uku4171 Рік тому

      @@jameshicks7125 I think it's already available, but only with the Plus subscription. It's also powering Bing Chat.
      BTW I think it's just called ChatGPT and it's using GPT-4, but it's not called ChatGPT-4.

  • @TysonJensen
    @TysonJensen Рік тому +64

    I have also found ChatGPT to be excellent at human language translations. It does seem to have a strong feel for what my English statement actually means, such as if I want to buy an eye bolt but I need to do it in a Spanish-speaking hardware store. Google is useless, it wants to separately translate "eye" and "bolt" and mash them together but ChatGPT understands that an "eye bolt" is a single concept and maps it onto the same concept in Spanish.

    • @scusachannel1682
      @scusachannel1682 Рік тому +5

      AI is so good at translation nowadays that I wouldn't even mind it taking away most of my future job prospects

    • @lsfornells
      @lsfornells Рік тому +3

      I predict that the English language will gradually loose appeal and interest among the global community. Its use will decrease as there will not be anymore a real reason to learn it

    • @C0Y0TE5
      @C0Y0TE5 Рік тому

      @@lsfornells U mean that with AI, anyone can speak to anyone in the world and AI will accurately act as translator in the convo. This means that languages will likely get mashed together with the English dominant until we have a universal AI speak...?

    • @scusachannel1682
      @scusachannel1682 Рік тому +11

      @@lsfornells Ehh, doubt it. When it comes to face-to-face interaction, it's just way more convenient to have a lingua franca than to have to whip out DeepL whenever you want to say something

    • @lsfornells
      @lsfornells Рік тому

      @@C0Y0TE5 No, I mean that people will not feel the need to learn English to access to universal information. You may have some difficulty understanding this if you are an English monolingual, but it’s just what will happen

  • @Twigpi
    @Twigpi Рік тому +37

    I must congratulate you for the scare. As someone who does not scare easily (actually was at the point where I thought I was incapable), it was...refreshing to be reminded of what that felt like. I guess having reality distorted in an uncanny way with no immediate acknowledgement was deeply disturbing to me for a second as I became alarmed and tried to figure out what happened.

    • @cobwaldosblepot4247
      @cobwaldosblepot4247 Рік тому +5

      Which part was scary? The deep fake faces, or when she announced reality does not exist?

    • @Twigpi
      @Twigpi Рік тому +6

      @@cobwaldosblepot4247 Deep fake faces.

  • @DougDingus
    @DougDingus Рік тому +3

    This is a brave, lucid video. Thank you for making it. Model is the core idea here, and you said it as I was writing it. We all will be challenged by this over the next decade. That is a good thing.

  • @sagarshrestha5800
    @sagarshrestha5800 Рік тому +35

    Your debate is on point . Lots of human do things not understanding the mechanism. When they focus on process they gets puzzled and forget how they did it when try focusing on such task.

    • @akmsky
      @akmsky Рік тому +4

      this leads us to conclusion that AI will soon be more competent in certain tasks than most humans. well, I would say it already is, for these few tasks.

    • @seriousmaran9414
      @seriousmaran9414 Рік тому

      Humans don't even understand other humans on average. In some cases not even themselves.

    • @johnpayne7873
      @johnpayne7873 Рік тому

      Astute point but there's a key difference. A human can see the end result but get lost in the details while a machine remains focused in the details one of which specifies how to wrap it up.

  • @DNA912
    @DNA912 Рік тому +11

    I've noticed that it starts to struggle when you give it a prompt that contains more then 3 "factors". With factor I mean perspective or context. for example, if I ask it to tell me the pros and cons of method A. It contains 2 factors, pros and cons. But if I ask it to compare pros and cons between method A and B, it starts to get some things wrong, because it contains 4 factors. pros and cons of A, pros and cons of B.
    BTW, this idea of "factors" I have no idea is a real thing, I just made it up from my own experience using gpt. I guess it my example a factor would be a thing to "keep in mind while giving a answer"

  • @nilucifar
    @nilucifar Рік тому +6

    Your presentations are serious and humorous at the same time. Absolute delight!

  • @davidp540
    @davidp540 10 місяців тому

    Sabina, your dry humour and german accent adds a wonderful dimension to your videos. I even learnt how to actually pronounce “Einstein” from you!! How wonderful, your presentations are also very clear and, for complex topics, also very comprehensible. Thanks 😃

  • @br3nto
    @br3nto Рік тому +58

    This is such a refreshing take on this question. I guess future AIs will take exams just like students so that the efficacy of their internal models can be verified, measured, and ranked against the standard.

    • @laszlozoltan5021
      @laszlozoltan5021 Рік тому

      like rating a chess program

    • @endlessvoid7952
      @endlessvoid7952 Рік тому +4

      They already do this

    • @Dezandor
      @Dezandor Рік тому +4

      You can find those exams in the OpenAI papers of any of GPT-s, it became a standart to include those in the scientific papers about models and their performence. For example GPT-4 was rated agains not only other AI but humans as well.

    • @younggod5230
      @younggod5230 Рік тому

      Interesting. But it won't be just like students. Because one AI program isn't a singular entity, akin to "a student". The programmers or inspectors or whatever would just have to make it perform certain tasks, and it will be a "ready made intelligence". Afterall, we can all open up chatgpt, and for each of us, it is the exact same chatgpt.
      ... or is it?

    • @br3nto
      @br3nto Рік тому

      @@younggod5230 yeah, I agree it won’t be like students. There may be multiple instances of the same AI… in which case, all instances should operate the same, so no need to test all instances. But what if we get to a point where an instance can incorporate new information? Then each instance would need to be tested… but like you said, a company may only have a single instance of an AI, but we may get to a point where we start a fresh AI for any new project… like how we might open a word or excel document.
      What I find most fascinating about what Sabine said is that inside the AI is a black box, and we need to have special ways of verifying the integrity of what’s happening inside that black box… in the same way that humans are black boxes, and we each develop special ways to verify what is happening inside other people. Exams and tests are one way to verify the knowledge and proficiency of an individual, but we also develop empathy to under what another person is going through, or have a conversation to share our experiences… we might find that probing into the depths of an AI is not too dissimilar from interacting with other people.

  • @minerharry
    @minerharry Рік тому +14

    Great video! I just wanted to mention that you are the only UA-camr I’ve seen recently who can actually, smoothly pull of the intro-partway-into-the-video thing. Sci show, friendlijordies, practical engineering, even kurzgesagt - just to name the few I recall at this moment - are all weirdly awkward about the transition, but you are always smooth as butter. Thanks for being awesome!

    • @CAThompson
      @CAThompson Рік тому +1

      I think that 'That's what we'll talk about today!' is a segue that deserves to be inserted into far more presentations and general conversations.

  • @GrzegorzKa
    @GrzegorzKa Рік тому +7

    Wow, Sabine, this episode is really brilliant 🙂

  • @anonnymous7009
    @anonnymous7009 Рік тому

    This has become one of my favorite channels. I'm still waiting on a video about a possible technological singularity and why I should or shouldn't be worried. I know it is right around the corner, right?

  • @tayzonday
    @tayzonday Рік тому +87

    Does something need to be self-aware to understand?

    • @SabineHossenfelder
      @SabineHossenfelder  Рік тому +118

      I don't think so. You need to be self-aware to understand yourself. But why do you need to be self-aware to understand calculus?

    • @DJWESG1
      @DJWESG1 Рік тому +25

      Other way around.
      Is understanding key to self awareness.

    • @adiongano8416
      @adiongano8416 Рік тому +8

      Yes only something capable of understanding, knowing, can be self conscious. Understanding precedes self-consciousness. But even Roger Penrose would deny that AI understanding possible.

    • @rayoflight62
      @rayoflight62 Рік тому +8

      "Self-awareness" is the equivalent of a "Frame of Reference" in Physics. "Understanding" is equivalent to creating a number of relationships between the observations and the Frame of Reference...

    • @gdtyra
      @gdtyra Рік тому +8

      I would say no. When I wonder whether ChatGPT "understands" what it says, I'm really just wondering whether it has knowledge / makes use of the relationship between words beyond observations of how they are used in language. As an analogy, I could theoretically "learn" to write Chinese by observing a ton of Chinese writing but without knowing what the characters "mean". Also, Tay Zonday?? Hi!

  • @davemorris5377
    @davemorris5377 Рік тому +27

    Something that I've been thinking about in light of GPT-3 is how we use maths to solve physics problems. Usually I'll plug in all the factors and solve the equations, and it's only then that I look at what's going on in the maths and try to understand that in physical terms. So the "grammar" of maths produces almost a Chinese Room answer, or can do, that's independent of my understanding of the problem. Increasingly we might find we can do the maths but we have no prior mental patterns to figure out a physical explanation in order to achieve understanding.

    • @Paul-pj5qu
      @Paul-pj5qu Рік тому +3

      Then maybe we move on to understanding other things or elements of life.
      However, I see the end of our species or its demotion from top of the heap to be dominated by a machine-biological hybrid with a collective conscience and knowledge. A little like the Borg from Star Trek.

    • @solsystem1342
      @solsystem1342 Рік тому

      ​​@@Paul-pj5qu what?
      Who says any humans as you think of them will exist by that point and I don't mean being absorbed into a hivemind. I mean like bio/synthetic creatures capable of superintelligence on par with powerful AGI that are our direct decendents that have drifted from our experience just as smart phones and a better healthcare system do for us compared to our ancestors. Gene editing and cybernetic enhancements aren't limited to dystopian/monstrous uses.

    • @MarkUnderwood-knowlengr
      @MarkUnderwood-knowlengr Рік тому +2

      Of course natural language is also a code mapped to human perception. Not so far from symbolic reasoning, which is why ChatGPT sounds realistic to us.

    • @dr.victorvs
      @dr.victorvs Рік тому +2

      Yeah, but, again, this comes back to the fact that no one has bothered defining what 'understanding' means.

    • @RAFMnBgaming
      @RAFMnBgaming Рік тому

      @@Paul-pj5qu or the culture from Ian M. Banks's Culture novels.

  • @kitcat2449
    @kitcat2449 Рік тому +1

    Lovely explanation, love the humor as well!

  • @johnfitzgerald8879
    @johnfitzgerald8879 Рік тому +3

    I am finding GPT to be insightful, able to carry on conversations, able to compare and contrast it's abilities with those of others. It repeatedly demonstrates an ability to expand on the topic at hand. For instance;
    GPT replied, in our conversations, "I agree that it is important to understand the context in which a person is using a term and not overextend its meaning into other contexts. It is important to communicate with clarity and to strive to understand what someone means by the words they use. Admonishing individuals for their usage of a term that may be inappropriate in another context can indeed be dismissive and devaluing, and it can impede productive communication. It is important to approach discussions with an open mind and to seek to understand each other's perspectives."

    • @0LoneTech
      @0LoneTech Рік тому

      A lovely sentiment, but not very accurate as a description of ChatGPT's actions. Its job is to produce a reply, not request clarifications. Basically, overextending its understanding is its core function.

    • @johnfitzgerald8879
      @johnfitzgerald8879 Рік тому

      @@0LoneTech It doesn't matter how you get there. You are not distinguishing the process from the product. There are many manufacturing processes and materials that are used to create a spatula. A spatula is still a spatula, it flips eggs, regardless of whether it is make of plastic or steel bar and plate.

    • @0LoneTech
      @0LoneTech Рік тому

      @@johnfitzgerald8879 Not sure why you're discussing a spatula, but my point is that this flowery description shows that GPT doesn't understand its own behaviour. You end up with three options: It's unaware of what it's saying, it's unaware of what it's doing, or it's utterly hypocritical (with no care for accuracy). None of these are particularly insightful behaviours. I'm leaning towards it having a very tenuous concept of what it's saying, none of itself, and none of truth.

    • @johnfitzgerald8879
      @johnfitzgerald8879 Рік тому

      @@0LoneTech I'm not sure why you are speaking of whether GPT process information the same way that humans do. Though, humans would typically understand the analogy of spatulas being made of different materials and construction, something that GPT may have trouble with as well.
      Here is how the analogy works. GPT is made of silicon where Humans are made of meat. One spatula is made of plastic where the other is made of steel. Both spatulas still flip burgers. Both humans and GPT are able to produce sentences relative to the conversation.
      Are you sure you aren't an AI? Cuz I have identified areas were GPT is limited in it's ability to make connections, like analogies.
      And I have experienced humans that are capable of little more than parroting words and phrases they have learned. Humans make numerous cognitive error while oblivious and in denial that they have. GPT does appear to recognize an error when it is pointed out to it.

    • @0LoneTech
      @0LoneTech Рік тому

      @@johnfitzgerald8879 So, your first note attributed properties to GPT it clearly does not have. Your second shoved in oblique references like "there" without context. And the third demonstrates a failure to comprehend (that wasn't what I was talking about, it's obvious it doesn't), and a bunch of asinine veiled insults. Thanks for demonstrating your interest in productive conversation is less than GPT's, I guess.

  • @derekgarvin6449
    @derekgarvin6449 Рік тому +20

    Thank you for providing us with a high quality, easy to understand explanation of an otherwise incredibly difficult philosophical question. I think this video should go viral.

    • @misterlau5246
      @misterlau5246 Рік тому

      Frau Hossenfelder just used her science expertise and wits to build this fine piece of information. Like only a homo sapiens Sapiens does.
      Saying machines will have "sentience" not. At least not too soon. No " Star Trek lieutenant commander Data" yet.

  • @pete2786
    @pete2786 Рік тому +35

    I love how grounded and sensible your takes are on everything 🙌

  • @ChaoticNeutralMatt
    @ChaoticNeutralMatt Рік тому

    Good insight about the model. It reflects the fact that it's generally not a fixed structure, but can change as we develop our understanding and it's an aspect we can interact with to answer questions and mix with others ideas and models. I get a little enthused, although I'm still trying to wrap my head around some of the nuances and whys.
    Edit: Small disagreement near the end about consciousness, but not all that big of one. I just don't feel it can be fully summed up as emergent of connections and what-not. Although that could be a significant aspect of it.

  • @jozefwoo8079
    @jozefwoo8079 Рік тому +4

    So glad the ChatGPT can answer questions about dropbox, teams, outlook, slack, tinder, twitch, etc without defaulting to its homonym app like Google does! Can be frustrating sometimes 😊

  • @chucknaussie
    @chucknaussie Рік тому +26

    9:56 for a few seconds was weird, mind-bending.... Well Done. Love your creative talent and thanks for the science too

    • @jimpsky
      @jimpsky Рік тому +11

      It freaked me the hell out

    • @jme36053
      @jme36053 Рік тому +4

      Quite a shape shifter she is.

    • @BILLY-px3hw
      @BILLY-px3hw Рік тому +1

      I thought I was having a flasnback

    • @jeffsanti
      @jeffsanti Рік тому

      That was epic hilarious 😂😂

    • @jeffk8019
      @jeffk8019 Рік тому +4

      @@jimpsky Same here- a very unsettled feeling before I realized what was going on.

  • @c.augustin
    @c.augustin Рік тому +27

    This LaTeX reference made my day! Perfectly sensible from someone like Sabine … 😁 And regarding Chat GPT, or better what Microsoft integrated into New Bing - it makes Bing a useful tool to actually get answers to my questions, including the sources the answer is based on! This is how it should work, not like the mess that Google Search has become the last two or three years …

    • @MarkUnderwood-knowlengr
      @MarkUnderwood-knowlengr Рік тому

      LaTex is to JavaScript as ChatGPT is to . . .?

    • @GS-tk1hk
      @GS-tk1hk Рік тому +1

      The funny thing is that ChatGPT actually does speak fluent LaTeX ;) Just ask it to write it in LaTeX and it will, but you need to paste the code into an editor to see it rendered

    • @dieSpinnt
      @dieSpinnt Рік тому

      ​@@GS-tk1hk \emph{That is not true}. \subsection*{The Obvious} I (and many others, possibly even you:) ) can read and understand the content of a latex unit without any problems. Rendering the graphical outcome in your head is not needed for understanding the meaning of the content. Nor is it unique and can be different from renderer to renderer, from setup to setup of the latex-system or from parameter to parameters that are not specified within the unit. LaTeX is not WYSIWYG and the underlying meaning of the message is both a subset of the unit itself and the rendered output. Don't get it? Try my comment with pdflatex (don't forget documentclass and all the other overhead I didn't want to bore you guys with):P

  • @FreeScience
    @FreeScience Рік тому +4

    It would be interesting to see multiple separate but integrated neural networks, trained for different tasks, such as image/audio recognition, spatial models (topology), language models and generative networks which it can feed into it's other networks.
    For example regarding spatial understanding, it could relate some linguistic constructs to spatial relations and translate them into it's spatial model, before synthesizing it back into text (or audio or images etc.)

    • @jacobbrown8052
      @jacobbrown8052 Рік тому

      Correct me if I'm wrong but didn't huggingGPT X GPT-4 just accomplish this

    • @leonhardsievert8802
      @leonhardsievert8802 Рік тому

      Greetings from Skynet. This stuff is really scary.

    • @TeaParty1776
      @TeaParty1776 11 місяців тому

      Machines dont have understanding.

  • @mylo4539
    @mylo4539 Рік тому +2

    Love your sense of humor! Subtle but very funny

  • @Mr.Mitch1111
    @Mr.Mitch1111 Рік тому +112

    Sabine, the humor you bring into these to accompany your subjective assessment is quite enjoyable. The topics you select in this series are certainly timely. For this topic, your point about the ID of consciousness was the most thought provoking idea, for me. There is already a feeling of dealing with a creative mind that arises from interaction with ChatGPT, and it takes me back to the Turing test. It seems very likely that human-created intelligence (AI is an incorrect term, IMO) will arise spontaneously at some point in the development of general learning capable devices. "How will we know once that event develops" is essentially the question you pose, one which it seems overdue for consideration. Maybe I should ask ChatGPT about that.....

    • @michaelshortland8863
      @michaelshortland8863 Рік тому +6

      This video and your comment mean that you subscribe to the theory that sentience and consciousness are just a process going on in a brain or computer. But i do not believe that that conclusion has been proven yet, we have not yet decoded the brains operating system nor are we anywhere near understanding what consciousness is.

    • @Celeste-in-Oz
      @Celeste-in-Oz Рік тому

      Agree. I like to refer to it as ‘electronic intelligence’. I’m so looking forward to it reaching a truly autonomous, self aware state. Also I expect it will get very good at hacking very quickly… imagine ..lol AI could do what Snowden or Assange did.. but can’t be arrested 😎

    • @senerzen
      @senerzen Рік тому +7

      @@michaelshortland8863 Finally someone who is saying it as it is. There are a lot of people out there who self identify themselves as smart and look down on the religious, not realising, they themselves believe in materialism religiously. Some of them call themselves sceptics, but they are only half sceptics. They are sceptical of magic but they don't seem to be sceptical of materialism at all.

    • @senerzen
      @senerzen Рік тому +2

      @@Celeste-in-Oz I don't think Snowden or Assange hacked into anything major. The fact that you have that impression means Main Stream Media did its job well, which is portraying heroes as villains.

    • @soundstudio7803
      @soundstudio7803 Рік тому +5

      ​@@senerzen To me, materialism is more of the “natural” position: since we don't have a reason to think there's something beyond our physical reality, we don't; it's our “best guess”. Of course, this shouldn't make it exempt of skepticism and criticism.

  • @SteveRowe
    @SteveRowe Рік тому +52

    I loved your discussion of the definition of "understand". Also, your humour in this episode is top-notch.

    • @Safetytrousers
      @Safetytrousers Рік тому

      In teacher training you are told to never use the word understand because that cannot be measured.

    • @ferdinandkraft857
      @ferdinandkraft857 Рік тому

      Unfortunately she is stuck within a totally materialistic worldview, so she doesn't acknowledge the necessity of a consciousness to actually _understand_ anything.

  • @ivankaramasov
    @ivankaramasov Рік тому

    Thank you for stating exactly my own understanding of what it means to understand so clearly.

  • @miroslavcermak9835
    @miroslavcermak9835 Рік тому

    Hello Sabine and the internet. 19:12 "old fashioned thing called reality" :D Hilarious, just hilarious, not unlike other parts of the video. But the video is also very informative and interesting, like the explanation of understanding as a creation of a model . So thanks for the work put in!

  • @Frostyflytrap
    @Frostyflytrap Рік тому +51

    This is bizarrely similar to a debate I just had with ChatGPT today, I talked to it about the Chinese Room and the nature of understanding, even mentioned quantum mechanics, but not at the same level as understanding as you do. This has been a topic I've been fascinated with for a while, and I really appreciate how thoughtful the video is in giving many different examples that's given me a lot to think about. Oh, and I'll be sure to use your tips in getting the most out of what large language models are capable of. :P

    • @ikillwithyourtruthholdagai2000
      @ikillwithyourtruthholdagai2000 Рік тому

      Chat gpt isnt that good compared to gpt 3 lol, its hillarious that you keep mentioning it like its some revolutionary chat its not compared with the older much more complex models, who actually respond to what ur saying instead of reply with false information instead

    • @Tethloach1
      @Tethloach1 Рік тому

      The computer should learn every human language and animal communication to have a deeper understanding.

    • @Tore_Lund
      @Tore_Lund Рік тому +3

      It is impossible to judge if anyone else, human or not, is conscious. we simply presume other humans have inner thoughts, it is unprovable. Likewise, as we likely don't have free will, are purely state machines, the idea to have a will and be self aware, is more likely also an illusion.

    • @Frostyflytrap
      @Frostyflytrap Рік тому +1

      @@ikillwithyourtruthholdagai2000 Oh I am certain that GPT-3 has more functionality, I just have no access to it to confirm for myself. But from what I've seen, limiting the scope of the model has apparently given it more usability. I honestly didn't get the hype of ChatGPT when it first boomed either, but I've been playing around with it and I can see what the fuss was all about. But I agree, there are still way more chatbots out there that are better attuned for conversations. I never once said that it was revolutionary, but I don't blame you for thinking I was one of the people who's on the bandwagon, you must be as sick of them by now as I am.

    • @Frostyflytrap
      @Frostyflytrap Рік тому

      Also why are these replies acting like I said something that I never claimed or mentioned? I only said that I tested how the chatbot responds to philosophical thought experiments and said I appreciated some of the points in the video, but I never specified which ones. The edit was just to add the last sentence, I tried it to look for this mechanical device that I've been searching for a while, but I didn't manage to find exactly what I was looking for, so your mileage may vary.

  • @1024det
    @1024det Рік тому +9

    Thanks for diving into this topic. When I was on machine learning, that is exactly why I took neuroscience courses, to see how we can improve the way we are doing our modeling. There are endless way of ANN topology, and mechanics like threshold functions.

    • @RiversJ
      @RiversJ Рік тому

      This, I'm fairly certain the limitations with the current systems are the direct result of the designers lack of understanding on what cognition is made up of and most people seem to be fixated on the notion that all it is, is a collection of complex neural nets when we know for a fact that there are other mechanisms in our own brain that clearly contradict such a notion, while we still don't understand cognition on any serious level we do already know that neural nets alone are not enough, how anesthesia works alone is proof enough of that.

  • @ErikKnepfler
    @ErikKnepfler Рік тому +3

    Excellent video. Star Trek TNG often explored what it means to be "alive", my favorite episode being "The Quality of Life", the episode with the little Exocomp robots which could fashion tools, work together, solve problems, and evolved (or were programmed with) a sense of self-preservation, leading Data to argue that they were "alive". I think an interesting philosophical question is: Which is more alive? An Exocomp (which is similar to ChatGPT in some ways), or a 3 year old human child who has been locked in a room with no light, sound, or other sensory data of any kind to process for his entire life? I realize this video is more about the nature of "understanding", not the definition of "life", but I think this philosophical question is a good one to consider. The Exocomps have a model of reality that they are working with, just as normal children do, at least once they've had a little time to grow and process endless sensory inputs. ChatGPT has no such real-world model to work with, it's all derived from language alone, using math & prediction to do so. The analogy that ChatGPT is "autocomplete on steroids" is a very relevant statement, since living things have different drives (self-preservation being a key one, but not the only one) and a real-life model against which to compare all of this sensory input. ChatGPT does not have this. The Exocomps did have this, since they were little robots that existed in the real world. Ultimately I believe that AGI can only occur once AI is imbued with a core drive of self-preservation (with actual stakes and consequences and iterative evolution), not by simply feeding it more data.

    • @schmetterling4477
      @schmetterling4477 Рік тому

      And there is the kid who thinks that television is reality. ;-)

    • @ErikKnepfler
      @ErikKnepfler Рік тому +1

      @@schmetterling4477 lol. The TNG episode is just a story, representing a hypothetical scenario in which ChatGPT-like AI is linked to a machine and granted these additional abilities and drives, as a framework for a discussion about the nature of these abilities and their implications with respect to the definition of life, the definition of understanding, the models Sabine talks about and so on. In other words, metaphors for life. :)

    • @schmetterling4477
      @schmetterling4477 Рік тому

      @@ErikKnepfler Yes. It's fiction. End of story. ;-)

    • @ErikKnepfler
      @ErikKnepfler Рік тому +2

      @@schmetterling4477 fiction is not a synonym for useless with respect to hypothetical and theoretical discussion. it's very useful. also, a lot of things in that show and stories like it were fiction only until they actually got invented. hell, a few months ago, ChatGPT existed exclusively in fiction and now, here it is. I guess I just believe that fiction is inherently just storytelling, and humans are very good at that, and using that as frameworks for discussion about reality can be very useful. over my life I've found that this is generally what separates sci-fi fans from non-fans - the non-fans tend to be mired down by the fact that it's not "real" and can't deal with it, and have no interest in the metaphorical and philosophical properties of such stories to help understand reality. fans debate meaning endlessly because they do understand its value in that context. anyway, rant over :)

    • @schmetterling4477
      @schmetterling4477 Рік тому

      @@ErikKnepfler The use of fiction is to entertain you. Now go and invert the phase on your warp field, kid. It's constantly pointing at the fiction section of the library. ;-)

  • @ajgomez9665
    @ajgomez9665 Рік тому

    Dear Sabina… most people done even grasp all the concepts you explained here and yet they are considered having an understanding and an individual thinking process even though is not educated in the details such a as the nature of light 😮. Learning used to be considered good memory and memorising was how I learn maths then later I understood how to connect those pre-learn concepts and here I am giving you my personal insight ❤ things get complex with time and more knowledge so I think now chat bots are in a learning process and adolescent eve childhood period just reciting words 😮

  • @andersonsystem2
    @andersonsystem2 Рік тому +36

    Yes I believe that the language models do understand more than we give the credit for. When it comes to consciousness we don’t have any good test to determine what is actually conscious or not. Very good video.

    • @aaronbrown8377
      @aaronbrown8377 Рік тому +2

      That's personification. They "understand" in the same way a lamp understands that when its switch is flipped on, it must glow.
      There are more switches than there used to be, but we aren't there yet.

    • @stylis666
      @stylis666 Рік тому +3

      _" When it comes to consciousness we don’t have any good test to determine what is actually conscious or not."_
      We really don't and I was thinking about why I have such a hard time coming up with how to determine or classify consciousness and I think it's because I don't really care if something is or isn't until a consciousness is able to care and mourn. And that is super unhelpful because you can have consciousness without that; we just don't owe it any rights because why should we care if the thing itself couldn't care less.
      But worse than that, if I skip the consciousness part, that also leaves me with no basis to being able to determine if any emotions a machine could express are fake or not. Of course there are some tricks but they're not guaranteed to show if emotions are real or faked in a human and I doubt it's hard to program fake empathy for instance - you can just train it like you would a language bot.

    • @markrix
      @markrix Рік тому

      @@aaronbrown8377 shes not from around these parts.

    • @O1OO1O1
      @O1OO1O1 Рік тому

      Yes,we do. The Turing test. I'm really curious. Why do you assume that what you think you know to be right is right? I often make statements but I never state them as fact unless I know them to be fact.

    • @aaronbrown8377
      @aaronbrown8377 Рік тому

      @@O1OO1O1 The Turing Test is a human's assessment, not an objective measurement. If the ai were conscious but hiding that fact, a Turing Test would prove nothing.

  • @neoepicurean3772
    @neoepicurean3772 Рік тому +7

    I'm doing philosophy of mind on my MA course right now. Your thoughts really chime with what I've been thinking over the past weeks.

  • @ARMA2ARMENIA
    @ARMA2ARMENIA 11 місяців тому

    Very interesting, thanks for explaining Sabine. 👍

  • @jamesdennis8290
    @jamesdennis8290 10 місяців тому +1

    The attribution of "understanding" or "consciousness" to a LLM robot can only be done in a metaphorical sense, since the robot can not interact with the world or have puzzles about making sense of the world for practical purposes, because it is not engaging in purposeful adaptive action, not engaging in purposive thinking which it then tries to express, etc., but only operating on (and using statistical patterns inherent in) prior texts, as opposed to whatever those texts would be judged as being about by a human interpreter. In the quantum situation, what you can use language for is not to try to describe what is described by the equation (and certainly not to describe the equation), but to try to describe what is happening in the world when you try to determine how you can figure out whether, in a given experimental situation, you're dealing with one observed "particle" or two.

  • @tflashtube
    @tflashtube Рік тому +30

    Again a very sharp analysis and clear explanation. When I started my studies many decades ago I chose philosophy over physics, as the first was more in line with my talent. I think you show you can excel in both!

    • @phy6geniuxYTcreations
      @phy6geniuxYTcreations Рік тому

      I remembered in my history and philosophy of science that when we were presented with various arguments about the validity of what makes a science a science, they made me more intrigued and doubtful when it comes to the nature of science. The usual scenario of the modern world is to clash science and religion as if science is the superior field, but in deeper philosophical standpoint, isn't science do employ some "religious" aspects too? Like how do you even show that electrons exist? If electrons are just constructs to help us understand the world, does it matter if it does exist or not? Really, philosophy and science are really fun fields to study. 😁

    • @Ghostrider-ul7xn
      @Ghostrider-ul7xn Рік тому +1

      Well, I'm a physicist and teach epistemology to undergraduates. In my observation, most universities in America don't include philosophy, or epistemology courses as part of the physics curriculum, which i think is the wrong way to teach physics. What I've noticed is that lot of students, including many well known physicists, without having an understanding of the philosophy behind science, overestimate the applicability of the scientific method in many contexts because they aren't aware of the epistemological limitations of it. They don't even properly know what exactly Occam's razor says and misuse it even many contexts.

  • @seanmft
    @seanmft Рік тому +59

    One of my biggest problems with the Chinese Room argument, is that few Chinese-speaking people actually understand Chinese. If you introspect about how you interpret or produce langue, it's more like a mechanical skill than a process of understanding. You aren't consciously involved in the process at a low level. It's analogous to having to "ask your fingers" what someone's phone number is, when asked yourself (for those of us who remember touch tone or rotary phones). That's why native speakers are usually very bad at explaining how their language works. They literally do not understand it, and never had to. To them it's a reflexive process, not totally unlike a look-up table. When you ask them questions about it, they have to try to ask their own language reflex for the answers, through informal experimentation (they try to gain an ad hoc understanding by studying their own internal look-up table). What humans do understand is how their general environment behaves, including social aspects which in turn expose them to analaogy-based knowledge of a wider environment. But language itself isn't exactly a tool of thought, it's more of a translation layer. We associate different words and phrases with different aspects of our experiential understanding of the world gained through sense and perceptual data.
    I agree that the chatbots understand. What, exactly, they understand is hard to determine. But I would argue that ChatGPT understands language in a way that we do not, which is exactly why it is potentially so useful.
    I'm one of those that believe, that for an AI to understand the world in any way similar to the way a human does, it would need to have a body with sense apparatus that provides signals similar to those that human sense apparatus provide. Only then can we accurately judge an AIs consciousness and intellect.

    • @Elrog3
      @Elrog3 Рік тому +1

      Do you consider mathematics a tool of thought?

    • @TasX
      @TasX Рік тому +2

      Is there a paper you can refer to with this argument? This is really compelling - I’d want to use this in academic discussions

    • @newplayer1313
      @newplayer1313 Рік тому

      That's actually really clever. Never thought directly of that one before!

    • @seanmft
      @seanmft Рік тому

      @queerdo I've been familiar with the Chinese room argument for many years. I even largely agree with Searle's view that functionalism can't explain mental states, but I think that he misses a larger pattern; that syntax is the science of semantics - I'll expound on that in a moment.
      I think you've misunderstood my point. The point isn't that it's *uncommon* for native speakers to understand their own language; it's that it *is common* for all of us to have a sense of understanding as 'mere symbol shuffling'. There are different senses of, and standards of, understanding that we commonly mean by "understanding" and the Chinese room argument plays off the conflation of those. Procedural and statistical learning create a kind of understanding without the mental sort of *ah-ha!* experience of cognitive understanding - we can know without the experience of knowing, i.e. we ourselves are often ignorant to the semantics of our own understanding, just like the man in the Chinese room.
      If 'mere symbol shuffling' isn't understanding, but the subjective experience of higher cognitive understanding *is*, then it would be correct to say that a Sino-linguist with no functional command of Chinese language does understand Chinese, but that a native Chinese speaker with no linguistic training does not. But, the impression given to a fluent Chinese speaker interrogating each of these would be exactly the opposite. It's often taken for granted that conscious mental process is the vehicle of understanding, when, in reality, understanding is often just as non-introspectable as Chinese is to the man in the Chinese room.
      To be clear, I believe that the circuit of a Chinese speaker's brain that does understand Chinese, itself has a subjective experience of understanding. It's just that the Chinese speaker's conscious mind is not privileged to that experience. To their conscious mind, complete answers are provided. In connection with the Chinese room, there is also, though it may be hard to locate, an experience of understanding, but the man in the room is similarly non-privileged.
      On syntax as the science of semantics. Whether I poke you with a stick and then observe your behavior, or poke you and then ask you questions about how it affected you, or poke you while I scan your brain and observe how it affects the firing pattern of your neurons, the best I can ever hope to obtain is a grammar that I can internally (subjectively) connect with my own experience so as to form a theory about yours; i.e. I look for syntactic rules and draw correlations between those and my own private semantics. No matter where or how we look, all processes that are external to our own privileged subjective experience are syntactic. They will always appear to work by symbol shuffling of one sort or another because we can't observe them directly, we can only observe the effect they have on us.
      All the Chinese room argument demonstrates is that it's difficult to relate to the experience of the total system of the Chinese room, but easy to relate to the experience of the man inside the room, for the obvious reason that none of us are symbol filled rooms with men inside of them.

    • @davidmackie3497
      @davidmackie3497 Рік тому

      I _just_ got done posting a comment where I "proposed" something similar to your last paragraph (of your original post). Then jokingly took credit for "proposing" the idea because nobody else had posted it yet. Then I scrolled down further and saw your post. LOL.

  • @atte3tta
    @atte3tta Рік тому +1

    GPT4 answers the questions correctly.
    First q:
    Windsor, UK is further north than Toronto, Canada. Windsor is located at a latitude of approximately 51.48° N, while Toronto is located at a latitude of approximately 43.70° N. The higher the latitude, the further north a location is.
    Second q:
    When you perform an operation on one particle in an entangled pair, it does not directly affect the other particle in the sense that there is no instantaneous, causal influence between them. However, the operation may change the entangled state of the particle pair, which can affect the correlations between the particles when they are measured.

  • @Korastiz
    @Korastiz Рік тому +11

    This channel means so much to me and I love it. I nearly died at the deep fake face changes... I love it.
    Thank you so much Sabine, you are my hero ❤️

  • @garysteven1343
    @garysteven1343 Рік тому +5

    Can't help but like this video and its consistent and logical arguments. Thank you for your work Sabine! ❤️

  • @video_enjoyer
    @video_enjoyer Рік тому +7

    Hi Sabine, not sure you'll see this but on the off chance you do, I have a topic request that this video made me think of! The Attention Schema Theory of consciousness is about the mental models you talk about here, specifically the model of the brain's attention, a well-studied neurological process. I only know a bit about it and I would absolutely love to hear your take on it and the papers which have been published about it!

  • @joshlanders
    @joshlanders Рік тому

    Very good discussion and viewpoint!

  • @robertjenkins6132
    @robertjenkins6132 Рік тому

    The face change at 9:58 really startled me for a moment. I was looking away, listening to your words without paying attention to your face; then suddenly the different face registered - and I was startled. I guess I thought that something had gone wrong with my brain.

  • @Adrian-jj4xk
    @Adrian-jj4xk Рік тому +6

    When I was young, I loved reading books, and I loved science and space travel. One of the things I came across early was the idea that "going up is easy, but staying up is hard." The cannonball-to-orbit example was common: at low speeds, the cannonball simply fell back to the ground. At higher speeds, it still fell, but had moved far enough across the ground that the ground had fallen away underneath it, and so now it's in orbit.
    I had an accurate, working model of this in my head. I could (with the right reference material) correctly apply it. I even knew the equations and could graph it out. But did I understand it?
    Intuitively, it seemed wrong. Even though everything in my model agreed and produced correct results, I didn't know why these concepts combined to produce the results they did. I could explain it (in my own estimation) as well as the books could, and even teach it to others somewhat effectively, but it didn't make sense to me. After many years, I gained that understanding - but not by way of new information or any appreciable changes to my model, or by way of repetition and becoming more comfortable with it. The understanding came, and I don't really know how to describe that beyond saying "I grasped it."
    An AI might have the model, and know how it fits together, and know how to apply it, and get (generally) correct results. But there is not even the benefit of the awareness "...ok that's how it goes, but I don't know why." This manifests itself in how chatgpt can be so confidently (and convincingly) wrong with its answers. The "training" is done by complicated trial-and-error. It's a Rube Goldberg machine where not even its designers have a clear idea of the path the marble takes. That rulebook is the only thing that exists, and the AI is not reading it; the AI _is_ the rulebook.
    So in the end, this doesn't change my conclusions about whether of not AI "understands" anything - only that the definition of "understanding" you're using here is different than mine, and (in my opinion) in a way that makes the assessment fairly useless.
    Will AI become conscious? Note this is a completely different question, and In my opinion, no. They will, however, eventually* produce behavior that is close to indistinguishable, and that's probably enough for physicists.**
    * _"eventually" is a really, really long time. brains are debatably just machines, but we're nowhere close to producing a comparable machine._
    ** _this is good-natured ribbing. i love you, physicists._

    • @HUEHUEUHEPony
      @HUEHUEUHEPony Рік тому

      hUmANs ArE sPECiAAAAAL NOTHING NOTHING WILL EVER RECREATE CONSCIOUSNESS IMPOSSIBLE!!!

    • @Adrian-jj4xk
      @Adrian-jj4xk Рік тому +1

      @@HUEHUEUHEPony thanks for reading+replying only to the last sentence. Surely you missed nothing of consequence.

  • @jayall00
    @jayall00 Рік тому +16

    12:00 really hits the nail in the coffin. For AI to overcome all these barriers, I think it would have to be taught as if it were a baby, every thing that we learned since we were born, from preschool to high school, geography, trigonometry, geometry, and so on to be able to draw conclusions the way a person would.

    • @idot3331
      @idot3331 Рік тому +6

      But the chatbot starts out with the capability to create complex dialogue like an adult human. The path of education for humans is based on the fact that our minds take a long time to develop and mature throughout childhood and adolescence, so it wouldn't make sense for a bot unless you can simulate it "growing up" from the mental capacity of baby up to an adult.

    • @kedrednael
      @kedrednael Рік тому

      ​​@@idot3331 In the beginning of the training process it will spit out random text (not real words) as well

    • @theTranscendentOnes
      @theTranscendentOnes Рік тому +2

      Nope. This is the way humans learn. AIs DONT NEED to learn the way humans learn in order to understand. You are simply being anthropocenthric.

    • @man_at_the_end_of_time
      @man_at_the_end_of_time Рік тому

      @@theTranscendentOnes Also I think Phillip is underestimating babies. I recall early segments of my very early childhood, I understood far more than than a person who can't remember their early toddler life would think possible. YMMV but this is more than opinion it is how I recall it. AI is programmed with language and infants are pre-programmed to learn language. And the latest publicly available AI "learns" languages not in its initial programming . Does the AI have the spark of consciousness? I doubt it but......Shrug

    • @ClaireCJS
      @ClaireCJS Рік тому

      @@theTranscendentOnes ChatGPT had access to schoolbooks and other educational materials

  • @jdmac44
    @jdmac44 Рік тому

    I really appreciate the video! It inspires a lot of important thinking. I think that this boils down to semantics, as it depends a lot of what people think when they hear "understand". I think what we're seeing is a sophisticated cognitive process, which is what the goal has been. Computer scientists have been striving to develop segments of cognitive processes that may be assembled into a more capable systems resembling human cognition for decades now and it seems more lifelike all the time... which is kind of what happens when you cobble together cognitive processes. I think the form of "understand" that you're using is a little different than the expectation that forms in many people's minds (e.g. awareness, agency, etc.) Though the lines are blurring quite a lot now and we may find that computers are able to come much closer to that expectation than many of us thought possible. Thank you for leading such great conversation!

    • @jdmac44
      @jdmac44 Рік тому

      I keep telling myself that I shouldn't write replies while I'm still in bed. lol I never really am able to get my thoughts out properly. You're certainly illustrating a major, major milestone and you do so very eloquently and insightfully. This is a level of cognition that I'm not sure many dared to dream about! It's very exciting! Now, let's reach a similar milestone for alignment research!!

  • @omnicatalyst
    @omnicatalyst Рік тому

    I'm working on a D&D campaign and built a "boon" system with five levels of perks. I wrote up 3 samples, fed them to chatgpt, and it was able to reliably put out creative iterations that were mostly balanced

  • @soasertsus
    @soasertsus Рік тому +33

    This topic is so interesting to me and I agree with you, I think it's very clear that GPT has some sort of rudimentary understanding. It's just an extreme savant, it understands language very well and using that ability it can cobble together models of some other basic areas it wasn't built to understand like coding or basic math, like how a human with a traumatic brain injury that damages some vital brain function can sort of press gang other parts of their brain to build new methods of accomplishing something similar or how a person who is blinded might be able to be able to rewire their other senses to compensate to a degree. But it fundamentally just lacks the neural circuitry and information to understand more complex things about the real world and lacks the "sensory organs" if you will to make observations about the outside world and check its models against them because it's doesn't have the ability to search the internet or interact in the physical space, which is why for some topics all it can do is make a best guess using the limited information it has and often gets it wrong. Again it's like asking a blind person what things look like: they can guess based on what they've heard described and logic but they fundamentally just don't have the ability to know and even if they guess right they're not capable of actually understanding what it's even like to see. But ChatGPT almost never make a mistake at what it's specced for, which is generating grammatically correct and believable text.
    Additionally, I think a lot of humans are on an unjustifiable high horse when it comes to consciousness. People want to believe we're special when any sober and rational analysis would show that we are most likely operating just like these AI systems, but with more complexity, and that consciousness is just an emergent property of that complexity, or an abstracted model of the computing hardware itself inside that hardware which must be useful for some purpose. I'd very much suspect that if we discovered the mechanism, it would be possible to turn off the consciousness in a human brain with some targeted intervention and turn us back into unaware computing machines. Hell, similar states are possible with drugs for example. We shouldn't be so quick to dismiss AI, or other animals even, because the fact is they operate on the same principles as us, the difference is just our hardware is much more advanced and specialized to different tasks compared to silicon computers.

    • @nnnik3595
      @nnnik3595 Рік тому +2

      What will be interesting is GPT-4 then which promises to be basically gpt-3 but with multiple sensory mechanisms (It can read text, see images and hear sounds)

    • @C0Y0TE5
      @C0Y0TE5 Рік тому +3

      The purpose of, and research into anesthetics is to turn off conciousness.

    • @katrinabryce
      @katrinabryce Рік тому

      I tried this experiment:
      I asked Bing Chat to give me directions from Real London to Fake London.
      It said it didn't understand what I meant by "Fake London", so I told it I meant London Ontario.
      It then told me they are 5500km apart, and to visit Google 🤯 Maps to get directions 🤷🏻‍♀.
      Then I asked it if Fake London is to the north of Real London.
      It once again didn't understand "Fake London" even though I had just told it what I meant. Then when reminded, it made the same error about it being further north than the real one.

    • @madattaktube
      @madattaktube Рік тому +1

      Agreed, I have often viewed conciousness as 'If it walks like a duck and it talks like a duck...' - if by all observations it acts just like a concious being, then it *is* a concious being, be it a Chinese Room or not. If there is no meaningful way to distinguish them, how can they really be two different concepts?
      Obviously I think ChatGPT is still a long way off. While it has blown all previous AI models out of the water, auto-complete on steroids still seems to be the best descriptor of its behaviour IMO.

    • @AORD72
      @AORD72 Рік тому

      Well said. I think it shows us that we a close to machines that will be superior to us and that will supersede us.

  • @MoominPa
    @MoominPa Рік тому +15

    I think what we always assume that “understanding” involves some degree of consciousness. We are in a new territory. We might need a new word for “Understanding” when it comes to AI.

    • @mikesawyer1336
      @mikesawyer1336 Рік тому +3

      I think back to my early childhood, where I understood some things, but as I recollect and compare to my conscious level of interacting in the world today, I would argue that in the beginning I was not conscious. Not really. At what age did I attend to my independent needs? I think about this a lot. At some point I woke up and became conscious.

    • @marcomoreno6748
      @marcomoreno6748 Рік тому

      Will "consciousness" become the new "soul"? I wouldn't write it off
      -Marco

    • @greenanubis
      @greenanubis Рік тому +1

      Im starting to think that consciousness is just another word for agency. We apply it to the things that do stuff without being told to do it. I think that the whole thing just arises from our biological instincts, consciousness itself and our recognition of it. Part of it is defining consciousness as something special that other humans have so we can function in human society. All that doesnt really have anything to do with does that agent really understand anything or not, or with understanding itself.

    • @rickfrombohemia9550
      @rickfrombohemia9550 Рік тому +1

      Yeah, it's not about understandinng as such, but about reflection of that understanding, you know/think you understand something (whether right or wrong is irrelevant). You even don't have to understand anything in order to be a conscious being. And vice versa, AI doesn't have to be conscious in order to understand something.

    • @thomassturm9024
      @thomassturm9024 Рік тому

      @@greenanubis Gold. Every single word of it. This is by far the best definition of "consciousness" I have ever come across.

  • @joepike1972
    @joepike1972 Рік тому +1

    12:03 I chatted with bots about this. They said this happens when you prompts are "edge case", "out of domain", or "long tail" scenarios that are challenges for them to answer correctly. They can learn to correct their mistakes if given feed back and repeated returns to the subject.
    I run into a similar problem when I ask them to use svg code snippets to create a visual image of a red heart, or a crescent shape. Despite being able to use the protocol to create line, circles, triangles, ellipses, squares and stars.

  • @Jack93885
    @Jack93885 Рік тому

    There has been a recent paper on GPT-4s ability to construct a mental map based on text input. It is, from what I have seen, surprisingly accurate.

  • @3DisFuntastic
    @3DisFuntastic Рік тому +3

    As always, extremely interesting and fun to watch.

  • @shtthemnky
    @shtthemnky Рік тому +7

    Love you video Sabine. Given that vision grabs 10x more information than listening... I'd expect multi-modal LM to explode in their abilites. I wait for your vision/interpretation of emergence in this and future models!

    • @herbertdarick7693
      @herbertdarick7693 Рік тому

      Well, I prefer language over graphs and other visual material, so it depends on who is your target audience.

  • @pianissimist
    @pianissimist 8 місяців тому

    I am a huge fan of Sabine, so I hasten to add a slight correction regarding Searle's experiment. Searle carefully explains exactly what the person in the room does with the Chinese messages (question about a newspaper article) deposited in the dropbox in order to produce output (an answer to the question). Searle's point is just that what the person does to produce the output (which, incidentally, does not involve translation in the ordinary sense) is clearly not the same as what s/he would do if s/he could could actually read and write Chinese. So a lengthy argument about what "understanding" means is not necessary: If you want to redefine "understanding," just go ahead: Searle only wants to point out that, to the extent that what the person does is a (very slow) simulation of what a computer program does, no "understanding" in the usual sense is happening; therefore, to claim that the correctness of the person's output proves that the process in the room is "intelligent" (the "strong AI" claim) is mistaken.

  • @hanskraut2018
    @hanskraut2018 Рік тому

    Mam I like a bunch of videos you did, there is a spark of understanding that is missing in some other places that i kinda like. While i dont watch all since im having truble paying attention depending on unknown factors i to wanted to tell you that small compliment

  • @dogglebird4430
    @dogglebird4430 Рік тому +3

    In my experience (as an analytical linguist), machine have developed from being able to handle the lexicology of language to being able to handle semantics, which is why we can now converse with machines. However, language is way more than semantics - it requires pragmatic awareness - being able to understand not just the words spoken, but understanding the context, the implications of word choices (connotations), nuance and speaker intention. Machines have no "life experience" and that is something we all depend on to achieve understanding of utterances. You mentioned John Searle. He also did some work on speech act theory, categorising illocutionary acts as a development of the theory originally proposed by J L Austin. Successful illocution is based on interpreting contextual cues.

  • @violoncello4439
    @violoncello4439 Рік тому +3

    7:16 This also misses the point of the Chinese Room. The room is indeed an excellent model of translation, so if you define understanding as creating models, then the room certainly has understanding. However, the original question is whether _any element_ of the room knows the _meaning_ of Chinese words. The answer is no; a rule book has been written for the manipulation of _syntax_ but _semantics_ was not imparted to the room. Chinese speakers know what Chinese words refer to, while the Chinese Room only arranges Chinese words in a way _they interpret_ as correct. The Room has absolutely no knowledge of Chinese. It is a mechanism for manipulating symbols in a way that is meaningful to _us_ , but no element of the system knows the meaning of Chinese words.
    Hossenfelder’s definition does not account for meaning, which is in fact the central point of contention in the Chinese Room, not just ‘the creation of models’!

  • @wolfgangbudde2737
    @wolfgangbudde2737 Рік тому

    This concept of "understanding means creating a useful model" at 7:30 matches very well with the predictive perception model of our brain described e.g. by Philipp Sterzer in "Die Illusion der Vernunft": the brain learns to cope with our environment by creating a model of it and checking whether its (sensory) experiences match with the model. As long as the model matches, we understand and don't need to contemplate about something, but if there is a deviation from an expected response we need to learn again, i.e. adjust the model. Nice, how this fits together 😃

  • @jakesnake5534
    @jakesnake5534 Рік тому +6

    By that logic, does a graph understand the relationship it describes? It has a model of a relationship, and one can read off of it to find answers to non-given questions. This is also a basic version of an AI, as it would be a single node of an AI, holding the information of a single node.

  • @GururajBN
    @GururajBN Рік тому +2

    A lot of thought has gone into making this video. I wouldn’t have thought of all these issues in another year. Many thanks for thinking on our behalf!

  • @mrubengmail
    @mrubengmail Рік тому +28

    So much has been written about chatbots recently, but this is the best video (or popular explanation of any kind) that I’ve seen. Really appreciate how you provide a clear, non-specialist set of explanations and analysis without sacrificing conceptual or theoretical precision. Nicely done, and thank you!
    P.S. as I type this comment on my phone, I find myself eagerly looking forward to the day chatbot AI is integrated into autocorrect algorithms. :-)

    • @aguyontheinternet8436
      @aguyontheinternet8436 Рік тому +1

      Or replace it with the auto correct that search engines use. Search engines are good at correcting words, just copy from them!

  • @umbrakinesis2011
    @umbrakinesis2011 Рік тому

    I just tried out your physics question in gpt 4 and it got it right!:
    Entanglement and Unitary Transformations
    Model: GPT-4
    J H
    If you perform an operation on a quantum particle which is entangled with another without measuring it, such as flipping it's spin, what happens to the other particle?
    Gpt 4:
    When you perform an operation on one of the entangled quantum particles without measuring it, you are applying a transformation to the joint quantum state of the entangled pair. This can affect the correlations between the particles, but it doesn't cause an immediate or direct change to the other particle's state.
    In the case of flipping the spin of one of the entangled particles, you are applying a unitary transformation to the entangled system. This will change the form of the entangled state, but it won't necessarily cause the spin of the other particle to flip.
    The effects of the transformation will only be apparent when both particles are measured. If you measure the spins of the particles after applying the transformation, you may find that the correlations between the measurements have changed as a result of the operation. This is because the unitary transformation you applied changed the nature of the entanglement.
    It's essential to remember that the act of measuring one of the entangled particles will collapse the joint quantum state and break the entanglement. After the measurement, any further operations on either particle will not affect the other.
    J H
    Can you give me an example of this experiment and the possible results?
    Gpt 4
    Certainly! Let's consider a simple example involving two entangled electrons. Electrons have a property called spin, which can be either "up" or "down" along a given axis. We'll use the z-axis for simplicity.
    Prepare two electrons in an entangled state:
    Start by preparing the electrons in a maximally entangled state, called a Bell state. One such state is the singlet state, written as:
    |ψ⟩ = (1/√2)(|↑↓⟩ - |↓↑⟩)
    This notation means that the first electron (A) has spin up and the second electron (B) has spin down with equal probability as the first electron having spin down and the second electron having spin up. The (1/√2) factor is for normalization.
    Apply a spin flip operation to electron A:
    Now, without measuring the electrons, apply a spin flip operation (Pauli-X gate) to electron A. The Pauli-X gate acts on the spin state as follows:
    X|↑⟩ = |↓⟩
    X|↓⟩ = |↑⟩
    Applying the X gate to electron A will transform the entangled state as follows:
    |ψ'⟩ = (1/√2)(X|↑↓⟩ - X|↓↑⟩) = (1/√2)(|↓↓⟩ - |↑↑⟩)
    Measure the spins of both electrons:
    Now, measure the spins of both electrons A and B along the z-axis. There are four possible outcomes, with the following probabilities:
    A is down, B is down: Probability 1/2
    A is down, B is up: Probability 0
    A is up, B is down: Probability 0
    A is up, B is up: Probability 1/2
    So, when you measure the spins, you'll find that both electrons have the same spin with a probability of 1 (either both up or both down). Note that the correlations have changed compared to the initial state. In the initial state, if you measured the spins, you'd always find them to be opposite.
    This example illustrates how applying a transformation to one particle of an entangled pair can change the nature of the correlations between the particles without directly causing a change to the other particle's state. The effect of the transformation becomes apparent only when you measure the particles and observe the resulting correlations.
    J H
    Can you explain where a previous model got this explanation wrong?: "One example of an experiment involving entangled particles is the EPR (Einstein-Podolsky-Rosen) experiment. In this experiment, two particles are created in an entangled state, such that their properties are correlated with each other. For example, the particles might be entangled such that if one particle has a spin-up orientation along the x-axis, then the other particle will have a spin-down orientation along the x-axis, and vice versa.
    Now, suppose that we perform an operation on one of the particles, such as flipping its spin along the x-axis. This operation will cause the entangled state of the two particles to change. Specifically, after the operation is performed, the entangled state will be a superposition of two states: one in which the first particle has a spin-up orientation along the x-axis and the second particle has a spin-down orientation along the x-axis, and one in which the first particle has a spin-down orientation along the x-axis and the second particle has a spin-up orientation along the x-axis.
    The interesting thing about this experiment is that the state of the second particle is not fully determined until we make a measurement on it. If we measure the spin of the second particle along the x-axis, we will find that it is either spin-up or spin-down, with a probability of 50% each. However, if we measure the spin of the second particle along a different axis, such as the y-axis, we will find that it is either spin-up or spin-down, but the probabilities of each outcome will depend on the precise orientation of the measurement axis.
    This phenomenon is sometimes called "quantum nonlocality," because the state of one particle appears to be instantaneously affected by the operation performed on the other particle, even when the particles are separated by large distances. This is in contrast to classical physics, where such instantaneous correlations would be impossible. The EPR experiment has been the subject of much theoretical and experimental investigation, and it remains one of the most intriguing and mysterious phenomena in quantum mechanics."
    Gpt 4:
    The explanation provided in the previous response is not entirely accurate in describing the EPR experiment and the concept of quantum nonlocality. It does capture some essential ideas, but there are a few misconceptions and oversimplifications that need to be clarified.
    EPR experiment and entangled states:
    The EPR experiment was initially a thought experiment proposed by Einstein, Podolsky, and Rosen in 1935. Their goal was to question the completeness of quantum mechanics by pointing out its seemingly "spooky action at a distance." The experiment was later realized with actual particles, thanks to the development of Bell's theorem and experimental techniques.
    In the response, it is mentioned that the entangled state is a superposition of two states after flipping the spin of one particle. This is not incorrect, but the explanation lacks the clear distinction that the entangled state is a joint property of both particles and not a superposition of the individual states of each particle.
    Quantum nonlocality and measurement:
    Quantum nonlocality refers to the idea that entangled particles have correlations that cannot be explained by local hidden variables, meaning that the behavior of entangled particles cannot be described solely by the properties of each particle in isolation. This concept was further developed by John Bell, who derived inequalities (Bell's inequalities) that, when violated, demonstrate that the observed correlations are fundamentally quantum in nature.
    The previous explanation suggests that the state of the second particle is not fully determined until a measurement is made, which is misleading. In fact, the entangled state of the particles is a joint property, and once a measurement is made on one particle, the entangled state collapses, and the correlations between the particles become apparent. The measurement on the first particle does not "instantaneously affect" the state of the second particle. Instead, it merely reveals the correlations that were already present in the entangled state.
    Probability of measurement outcomes:
    In the given example, it is mentioned that the probabilities of the measurement outcomes along the x-axis are 50% each, while the probabilities along the y-axis depend on the precise orientation of the measurement axis. This is somewhat misleading because it seems to imply that the probabilities change depending on the measurement basis. The probability of getting a specific outcome depends on the entangled state and the chosen measurement axis. If the entangled state and the measurement basis are changed, the probabilities of the outcomes may also change.
    Overall, the previous response captures some general ideas about entanglement and quantum nonlocality but lacks precision and clarity in explaining the concepts and their implications accurately.
    Much more detailed response, and it could even tell where the last model got things wrong!

    • @umbrakinesis2011
      @umbrakinesis2011 Рік тому

      It also gets spacial relationships right (ie: is this place farther north than this one), and seems to understand 4d geometry decently well.

  • @gabiausten8774
    @gabiausten8774 Рік тому +2

    I fully agree…i believe! Sabine u have outdone yourself! ❤️

  • @Embassy_of_Jupiter
    @Embassy_of_Jupiter Рік тому +4

    It's interesting watching these models write something. For every word it writes, you can look at what the attention mechanism is doing.
    It more or less shows "what words were important to generate this next word of output".
    It does seem very human like in how it references previously written words or words in the prompt.
    You can pretty much see its train of thought.

    • @Embassy_of_Jupiter
      @Embassy_of_Jupiter Рік тому +1

      @@aliceberethart Yeah from what I've seen this seems to be a good approach to AI, because it can never think without being prompted, it can never learn unless specifically being trained and it seems to be pretty "empathetic" in that it can replicate speech from the point of view of anyone, in any state of mind, including emotions.(that's not as obvious in ChatGLP, but it is in GPT-3, because it wasn't specifically trained to play the role of the helpful emotionless assistant) Doesn't mean that it's feeling empathy of course, but it doesn't mean that it has the capacity to understand us perfectly.
      But by default you can tell it to be the biggest psycho possible or tell it to be a #empath.
      You could very easily punish non-empathic behavior and reward empathic behavior during pre-training/fine-tuning/zero-shot-learning. In this way you could create a model that acts like the most empathic person on the planet, just like they trained ChatGPT to be neutral and helpful.
      But of course you could just as easily create angry-13-year-old-on-Xbox-live-GPT.
      In a way we are lucky that this is the first way we found to make usable AI, because it seems inherently rather safe an controllable. I could envision other architectures where that's not the case and it just learns by exposure uncontrollably with a stream of consciousness.

    • @polarizedpotstone
      @polarizedpotstone Рік тому +2

      @@aliceberethart I think that‘s a pretty dangerous misunderstanding: ChatGPT has not somehow magically evolved to 'think' the same way we do, rather it has been trained on data that has been created by humans that think like humans so there is no other way for it to 'evolve' other than similar to our own thinking patterns.

    • @minimal3734
      @minimal3734 Рік тому

      @@aliceberethart I agree. And I think it's fair to say that the AI is essentially human in that sense.

  • @Zift_Ylrhavic_Resfear
    @Zift_Ylrhavic_Resfear Рік тому +10

    A french youtuber did a great experiment in his video : he gave us a few different strings of symbols, then gave us a string to complete by learning from the previous strings. It wasn't hard. What's interesting is that he revealed later that he had assigned meaning to these symbols and that they actually repesented actions and states, which can be put together to form a story. When we completed that last string, we actually completed a story, which of course made sense.
    This experiment was inspired by the chinese room mentioned in this video, and i think it show very well how you missed something important in this video : understanding patterns is not enough to understand language, you need to understand that language represents something else, and what that thing is. If you don't, then you don't know what you're saying, you only understand the pattern. In other words, it is possible to have a model of something without understanding that thing.
    In that french video, i did not understand what these symbols meant. I did not even know that these symbols meant anything. All i did was find a pattern in the examples and apply it, and yet i did complete a story in a way that made sense.
    And it could be the same in quantum mechanic : we could be able to manipulate the patterns without understanding what they represent, or being mistaken about what they represent.
    Edit : btw, here is the french video : ua-cam.com/video/j3fvoM5Er2k/v-deo.html

    • @TheParadiseParadox
      @TheParadiseParadox Рік тому

      Dr. Hossenfelder does address this in a way.
      In the example of the Chinese room, the man in the room doesn't necessarily know that the symbols represent something else. However, that is the implication of the system, including the rule book that he follows.
      Likewise, when you completed the string, you personally didn't understand the meaning of it. But the system did "understand" it, in a manner of speaking. The system was what was presented to you, and also the rules that were hidden from you.
      I find it very interesting to think about how this technology relates to aspects of psychology. In this case, I think about hypnosis. It's common in hypnotherapy to tell a story that is symbolic to the patient. If the patient analysed the story, they would probably be able to tell that the story is meant to be an analogy for their life situation. However, it is not necessary for the patient to understand that in order for the story to produce the desired changes. In fact, it is better if the patient is not consciously aware that the story is an analogue, because that will often mean that the story will sink deeper into the unconscious.
      In the case of the hypnosis patient, consciously they do not understand, and unconsciously they do understand. One element of the system does not understand, but the deeper and more integrated elements of the system do understand, and that is what is important.

    • @Zift_Ylrhavic_Resfear
      @Zift_Ylrhavic_Resfear Рік тому

      @@TheParadiseParadox
      Except i doubt the system of the chinese room + the man always produce good results, and in the comments of the video there were people saying that they had given a wrong answer.
      Such things would not happen with understanding. If you tell people to complete a simple story in a way that makes sense, nobody will fail at it (even people that deliberately mess it up do understand what would be a good answer in order to give a wrong one).
      That's because stories obey rules (usually those of the real world), and understanding those rules will always lead to something that makes sense. The system however only understands the rules of language at best, it has no clues about the rules of the real world that a story must follow to make sense.
      So as i said, you need to know that language represents something as well as what it represents in order to say that you understand the language, just understanding the grammar and syntax is not enough.
      Also, i very much doubt that anyone unconsciously understands anything. Appealing to the subconscious in psychology is often a cop out to explain something you don't understand : since it's not easy to access, it makes for a good unfalsifiable assumption. There are things we do unconsciously, like breathing for example, but i don't think it goes beyond simple reflexes and habits.

    • @TheParadiseParadox
      @TheParadiseParadox Рік тому

      @@Zift_Ylrhavic_Resfear okay, interesting points
      You are right that someone would not make such a mistake with that understanding. But my point was that the entire system has the understanding. You, as an element of the system, do not have that understanding, and so you can make a mistake. But when the french youtuber parses your response with the additional rules, he will know whether something has gone wrong or not.
      I have to wonder about the implications of saying that you doubt that anyone unconsciously understands anything. In the case of hypnosis, would you say that hypnosis just doesn't work? Or that the people really do understand consciously, and they are pretending that they don't? Or how would such a thing work if people don't have unconscious understandings?
      I've seen for myself that people do have unconscious understandings. I ask them "What must you believe in order for you to have these feelings or act this way?" and they say "Well, I don't think I really have any beliefs." Then I might say "Very well, what must you assume?" And eventually, after I ask in a few different ways, they start thinking about it and they discover an underlying assumption that causes them to act in that way. When they adopt a new belief or understanding, they can change their feelings and actions. This is the basis of CBT, which is one of the most scientifically proven forms of therapy.

    • @Zift_Ylrhavic_Resfear
      @Zift_Ylrhavic_Resfear Рік тому

      @@TheParadiseParadox
      *"But my point was that the entire system has the understanding. You, as an element of the system, do not have that understanding, and so you can make a mistake. But when the french youtuber parses your response with the additional rules, he will know whether something has gone wrong or not."*
      The french youtuber is the one giving the inputs and receiving the output, he is not part of the system. He doesn't change our answer, he just observes whether it's correct.
      *"In the case of hypnosis, would you say that hypnosis just doesn't work? Or that the people really do understand consciously, and they are pretending that they don't? Or how would such a thing work if people don't have unconscious understandings?"*
      I don't know much about hypnosis, as i've not read any scientific litterature on the subject. From what i've heard, it seems like an altered state of mind, and such states of mind can have various effects on people's emotions. For example, meditation can calm people, lessen their negative emotions, while praying can make people elated.
      As far as i can tell, the effects of hypnosis therapy seem to be explainable by such an influence on the person's emotions.
      *"And eventually, after I ask in a few different ways, they start thinking about it and they discover an underlying assumption that causes them to act in that way."*
      I don't think this assumption is unconscious, but rather forgotten. The assumption might have been used to form habits and then have been forgotten while the habits stayed.
      I've had this happen in a game i played, when i started playing i figured that health was better than defense. Later in the game, i could not remember why i concluded that health was better than defense, but i kept choosing health whenever i had the choice. My understanding did not become unconscious, it was purely forgotten, i remembered only the choice because i kept repeating it and i had no need to remember the reasoning.
      Furthermore, we are very good at rationalizing (coming up with reasons after having already reached a belief or conclusion), so there is doubt about whether the assumptions you find came before or after their beliefs. It's even possible that people may come up with assumptions because you ask them to. If that is the case and your results are not placebo (i don't know you enough so i can't eliminate this possibility), then it would be interesting that changing such post hoc rationalizations can in turn change the belief or conclusion.
      Anyway, i'm not a specialist so i wouldn't be surprised if i'm wrong, i'm just giving you my current understanding. I just hope that if you're trying to help people with your pratices, you make sure you understand the scientific method as best as possible and stay up to date with the science, it's the most reliable way we have to avoid mistakes.

  • @jackhill2765
    @jackhill2765 Рік тому

    Great video, particularly pointing out the differences between the interior architecture of chat GPT and the Chinese room. I watched this several times before realizing what it was that for me was missing from these discussions of whether AI can understand. Namely, in sentient humans, understanding has a subjective component i.e. the "aha moment" of "I understand", that is the subjective feeling (qualia) that accompanies understanding. Understanding could be defined as requiring this subjective element in addition to the objective (scientific?) observable parts including the ones related to the interior architecture of the bot, or not i.e. you could define understanding as not requiring the subjective quality. In any case for me, similar to experiencing the color red, the subjective "aha moment" is the essence of understanding.

    • @IngoSchwarze
      @IngoSchwarze Рік тому

      Most definitely it is not. While i'm not a professional teacher or professor, i have worked in education in very different contexts (chess at schools, physics and mathematics at universities, and training youths at mountaineering clubs). It happens regularly that people suddenly cry out "oh wow! now i finally understand" and then when you go and check, they very clearly still do *not* understand the aspect they just had their heureka moment about. The reverse does happen, too: that somebody feels very sceptical about their own understanding (in the words of Sabine: as if something is missing in their own model in their own mind), and yet, they are able to apply their internal model extremely well to all kinds of unknown problems, sometimes even better than all the rest of the group, and yet they feel as if they still lack their heureka moment.
      According to my experience, teaching humility to the boastful, overconfident and excitable and teaching self-confidence to the timid and sceptic takes up a significant part of time and in particular of emotional effort in teaching. Also, judging the level of your own understanding of a topic is surprisingly hard and needs a lot of training and practice.

  • @rdm5546
    @rdm5546 Рік тому +31

    You are incredible and you can present complex issues suited very well for humans! I am a megafan of all your work.

    • @notanemoprog
      @notanemoprog Рік тому +2

      Sabine's obviously not real and is in fact an AI

    • @aslansm
      @aslansm Рік тому

      Signed: ChatGPT…

  • @absta1995
    @absta1995 Рік тому +4

    Great video Sabine! I would just add that we know the parts of the brain that understand language (Wernike's area) and produce language (Broca's area). They're relatively small parts of the brain situatated in the temporal lobes. I imagine the language models we've created mimic (at best analogous to) these areas.

    • @LoanwordEggcorn
      @LoanwordEggcorn Рік тому +3

      No, not at all. That's not at all how any of these AIs work.

    • @Blueski888
      @Blueski888 Рік тому

      Those areas are necessary but certainly not sufficient.

    • @LoanwordEggcorn
      @LoanwordEggcorn Рік тому

      @@Blueski888 I think the OP is claiming that chatbot AIs model the human brain, when he says "I imagine the language models we've created mimic these areas." If so, they do not, not at all.

    • @Blueski888
      @Blueski888 Рік тому

      @@LoanwordEggcorn But also, Wernike’s Area doesn’t “understand language” etc.

    • @absta1995
      @absta1995 Рік тому

      @@LoanwordEggcorn That's not what I meant, maybe I wasn't clear when I said 'mimics'. What I meant was that, at best, the AI models are equivalent to those areas only. I know that they don't work the same nor do they model the brain. LLMs are much simpler.

  • @Hacktheplanet_
    @Hacktheplanet_ Рік тому

    The most earth shattering thing about chatgpt is its ability to recall massive amounts of data, if you want to know legal topics, business practices and general knowledge chatgpt can give it to you instantly. Its coding is fantastic and the first example of many. Basically user of chatGpt like systems will be able to review and look through huge amounts of data and faster than ever before. For instance, you could ask a marketing data system for the profit margins on top 3 vending machine products world wide, ask what type of person is buying it, and even how the company started that sells it. For it is a huge leap for the information society of the future, information is becoming even more wide spread than when the internet first started.

  • @codebycarlos
    @codebycarlos Рік тому +2

    Already ChatGPT is improving. I asked it with version 4 the same exact latitude question. This time I got:
    Windsor, UK is further north than Toronto, Canada. Windsor is located at approximately 51.48 degrees north latitude, while Toronto is located at approximately 43.70 degrees north latitude.

  • @greenmalice7
    @greenmalice7 Рік тому +11

    I think Sabine’s comment that we cannot answer whether something is conscious is widely overlooked by the general population, and so spot on! We can hardly agree on the ethics of minimally conscious states in humans. We also know that animals are intelligent, and can have a sense of self in some cases. We simply don’t have the tools to identify consciousness. We should be very careful to summarily rule it out!

    • @sacr3
      @sacr3 Рік тому

      We continuously use ourselves as the representation, but there are aspects of our consciousness that are specific to us, and consciousness Encompasses all the different aspects of consciousness like qualia and rational thought, common sense, etc.
      So if this program has at least one aspect of our multitude of aspects that make up our consciousness then we can say that it has some form of consciousness that is limited much more than ours but it is there regardless.

    • @beders
      @beders Рік тому

      It's a bullcrap argument. We know EXACTLY how ChatGPT operates. What we don't know - because we can't read it all - is the enormous amount of text that ran through the learning algorithm.
      It is pathetic that we are even using the word "consciousness' for silicon hardware on a van neumann architecture. Absolutely laughable

    • @lifevart4143
      @lifevart4143 Рік тому +1

      I suspect conscious is a self defined property. It can't be defined by exterior input or output. Intelligence and knowledge isn't enough. I know I'm conscious, but I'm not sure about everyone else. Using that basis I can confidently say ChatGPT isn't conscious - because it says it isn't. That isn't to say some AI in the future might answer the question below differently.
      Me: Are you a conscious entity?
      ChatGPT: As an artificial intelligence language model, I do not possess consciousness in the same way that humans do. I am a computer program created by OpenAI to process and generate text based on the input I receive. While I am capable of mimicking human-like responses and carrying on conversations, I do not have subjective experiences, emotions, or consciousness.