Francois Chollet - LLMs won’t lead to AGI - $1,000,000 Prize to find true solution

Поділитися
Вставка
  • Опубліковано 8 січ 2025

КОМЕНТАРІ • 1,4 тис.

  • @cmfrtblynmb02
    @cmfrtblynmb02 7 місяців тому +503

    One advice to host: You need to give your guest space. You are not a salesman. Or a missionary. Challenging them does not mean repeating the same argument over and over again. It was suffocating to listen to your challenges. If it was not for the call and patient demeanor of Chollet, it would be impossible to watch. We were not able to listen to Chollet expanding upon his ideas because host just reversed the clock to zero by repeating the same "but memorization is intelligence" argument. It should be about your host, not showing the supremacy of your ideology or beliefs. If your host is wrong, you can prove them wrong by showing their arguments and ask questions as they expand upon them and then show if they are inconsistent. Not repeating the same thing over and over and over again

    • @mmelanoma
      @mmelanoma 6 місяців тому +25

      this. on the other hand, it gave me tons of reaction images hahaha, some of francois' sighs and nods are just gold

    • @dontwannabefound
      @dontwannabefound 6 місяців тому +7

      💯

    • @Limitless1717
      @Limitless1717 6 місяців тому +18

      One advice to you: You can make this point better without insulting Dwarhesh, who is a young man that is still learning. Perhaps you should try hosting a podcast and see if you do better. Want my guess? You would do much worse than you think.

    • @aga1nstall0dds
      @aga1nstall0dds 6 місяців тому +3

      ​@@Limitless1717ya ur right lol... he just a lil bit slow and no offence to him it just demonstrate how francois is a genius

    • @cmfrtblynmb02
      @cmfrtblynmb02 6 місяців тому +42

      @@Limitless1717 this is the most mundane criticism of a criticism. Dwarhesh himself is doing a podcast on topics he is not a specialist on and he is openly criticizing and challenging the views of a specialist on a topic here. So maybe he should work on AGI before challenging François here if he were to take your advice seriously (though he should try to educate himself on topics in any case)
      And I am not doing podcasts but I have taught many many classes with a lot of teaching awards. Not the same but similar when it comes to expanding on topics. When I teach a concept I don't just attack it on the first sentence. I explain it, allow it to ripe. Put different light on different aspects of the topic. I don't try to destroy the whole concept on the first sentence.
      So my advice doesn't come out of nowhere. And if he puts himself to public spotlight, my criticism is actually one of the most innocent stuff that he is thrown in his direction. If he takes into account he can improve upon. I am mostly criticizing how he is doing some stuff and even provide what he can do better. It is weird that you take offense to that. Anyways it is up to him to do what he wants but I won't watch him anytime sooner again. As it is now, this is really bad way of arguing with anyone even in private, let alone on a podcast. When someone interrupt me like he does and don't know how to argue, in general I just don't bother

  • @MM-uv3vb
    @MM-uv3vb 7 місяців тому +733

    Need more interviews with legit AI/AGI skeptics to balance out the channel

    • @mbican
      @mbican 7 місяців тому +7

      There is money in that

    • @benbork9835
      @benbork9835 7 місяців тому +101

      Skeptics? You mean realists

    • @danypell2517
      @danypell2517 7 місяців тому

      nah fck the doomers

    • @ModernCentrist
      @ModernCentrist 7 місяців тому +47

      The issue is, highly intelligent AI skeptics are in short supply.

    • @EdFormer
      @EdFormer 7 місяців тому +49

      ​@@ModernCentristthat's the impression you get from content focused on hype merchants, alarmists, and those with a vested interest in overinflating the value of current methods. Academia is full of skeptics/realists.

  • @BrianPeiris
    @BrianPeiris 7 місяців тому +66

    Really glad to see people like Chollet are willing to say the emperor has no clothes. I'd recommend Subbarao Kambhampati's talks as well. He goes into some theories about _why_ people are being fooled into thinking that LLMs can reason.

    • @jimgsewell
      @jimgsewell 7 місяців тому +6

      Thanks for the additional resource

    • @benprytherch9202
      @benprytherch9202 6 місяців тому +2

      Subbarao is wonderful! Great recommendation.

    • @therainman7777
      @therainman7777 10 днів тому +1

      I wonder if you’d still stand by these comments now that (1) o3 and o3-mini have both shown remarkable performance increases on ARC, despite both of them being LLMs (and claims that these models are NOT just LLMs are incorrect; OpenAI engineers themselves have confirmed that they are LLMs, just trained using additional techniques), and (2) Francois Chollet himself has now changed his tune in response to these new results, conceding that LLMs such as o3 and o3-mini “could be doing reasoning, as a process of iterative search for programs in natural language space.” IMO, while Chollet and Subbarao are both very smart and have some wonderful insights, they were both overly confident about the inability of LLMs to reason, and the empirical reality is now starting to show that. Chollet himself seems to be conceding this, at least in part. So I wonder how you’d react to your earlier “emperor has no clothes” claim.

    • @BrianPeiris
      @BrianPeiris 10 днів тому

      @@therainman7777
      Firstly, you have to trust me that I'm not looking to fight anyone on this, or stick to my own dogma. I've been trying to wrap my head around all of this like anyone else.
      Here are my points:
      1. o3's progress is undeniable. We've entered a new era with these inference-time solutions
      2. o3 is based on LLMs, and trained to have better abilities, but it is clearly not just an LLM. This is evident because you cannot get better results with LLMs by increasing inference-time compute. Where as the o3 solution does get better with more inference-time compute. Therefore it is a new paradigm.
      3. So yeah, I stand by these comments with regards to LLMs, but now that we've entered this new era, my comments are no longer important. The focus must naturally shift to inference-time solutions.
      4. Personally I'm always interested in the failure modes, because they reveal the most. o3 was unable to solve 9% of the ARC test set, even when it was given many thousands of dollars worth of inference-time compute. Some of the tests it failed are trivial to humans. Why is that? Is it really doing reasoning?
      5. Chollet claims the upcoming ARC-AGI-2 benchmark will confound even o3 at high-compute, while still being easy for humans. I look forward to that because, again, my desire is for true AGI, not hype.
      If you'd like to continue this conversation, I'd be happy to, but please let's do that anywhere other than youtube comments. I really dislike this reply format, and youtube just randomly eats the comments I post. You can find me at brianpeiris on discord or by email, brianpeiris AT gmail

    • @BrianPeiris
      @BrianPeiris 10 днів тому

      @@therainman7777 I'd like to give you a proper reply, but youtube keeps eating my comments. I'd be happy to continue the conversation but please let's do that anywhere else. I'm brianpeiris everywhere.

  • @therobotocracy
    @therobotocracy 7 місяців тому +574

    What I learned from this: Be the guy who speaks slower under pressure, not the guy who talks faster!

    • @kjoseph7766
      @kjoseph7766 7 місяців тому +7

      what pressure

    • @therobotocracy
      @therobotocracy 7 місяців тому +48

      @@kjoseph7766 They were clearly putting their own opinions forward, which didn’t align. I would say that put pressure on each other to defend their beliefs. The one dude got super panicked as the French guy kept putting forward clearly articulated ideas and rebuttals.

    • @ahpacific
      @ahpacific 7 місяців тому +17

      After listening to this interview and reflecting on it, I was actually thinking that (not Dwarkesh because he's legitimately smart in my opinion) but "people" like Destiny talk extremely fast to compensate for shallow ill-formed thoughts.

    • @Gome.o
      @Gome.o 7 місяців тому +17

      @@ahpacific Dwarkesh could help himself out a lot if he slowed down. Simple things like rephrasing what 'french guy' was saying in a way that Francois would agree with would also help tremendously. There is a fundamental difference in epistemology between these two, Francois is emphasising true understanding, and Dwarkesh seems to imply that large gross memorisation leads to understanding- which I don't think Francois would agree with

    • @wiczus6102
      @wiczus6102 7 місяців тому +2

      ​@@ahpacific I am pretty sure Destiny is very transparent on when his thoughts are shallow or not. Notice when he formulates things as open questions and or the adjectives and intonations he uses to suggest the difference between perspective and fact. You can make a false statement and still create a good converstation like that. People like that are fun to talk to as opposed to people who will only say something if they are fully certain.

  • @cj-ip3zh
    @cj-ip3zh 7 місяців тому +617

    Dwarkesh is LLM, Francois is AGI

    • @ClaudioMartella
      @ClaudioMartella 7 місяців тому +21

      perfectly said

    • @cj-ip3zh
      @cj-ip3zh 7 місяців тому +9

      @@ClaudioMartella I may have commented too quickly, it seems thar later on in the video I got the impression Dwarkesh was playing devils advocate. Not sure...

    • @wiczus6102
      @wiczus6102 7 місяців тому +15

      @@cj-ip3zh By impression you mean that he literally said he is playing the devils advocate?

    • @cj-ip3zh
      @cj-ip3zh 7 місяців тому

      @@wiczus6102 Right but I couldn't tell if he meant just in that shorter exchange or if the whole interview was him taking the opposite site for the debate.

    • @XShollaj
      @XShollaj 7 місяців тому +1

      Damn - perfectly put!

  • @perer005
    @perer005 7 місяців тому +436

    Dawarkech does not appear to have the ability to adapt on the fly in this interview! 😂

    • @vsanden
      @vsanden 7 місяців тому +34

      He showed by example that some humans are not AGI

    • @slm6873
      @slm6873 7 місяців тому +31

      One of the worst and least informed interviewers, the only way I can get through his videos is fast forwarding through his incessant low information babbling. Yet he keeps getting very smart people on to interview!

    • @mm100latests5
      @mm100latests5 6 місяців тому

      @@slm6873 dudes like this are good for pumping up the stock price

    • @skoto8219
      @skoto8219 6 місяців тому +2

      @@slm6873can you tell us what the dumbest question was and why it was dumb?

    • @IndiaEdition
      @IndiaEdition 6 місяців тому +5

      I was gonna like this comment, until I watched the whole talk.
      Guys watch the whole talk before forming an opinion!
      really interesting. @dwarkesh - great going as usual

  • @TheMrPopper69
    @TheMrPopper69 7 місяців тому +568

    Dwarkesh dismissed a few completely valid answers to try and steer the answer in his preconceived idea of LLMs, I didn’t like that, dude is smart, let him finish and actually take onboard his answer before asking another question

    • @therainman7777
      @therainman7777 7 місяців тому +32

      He said he was playing devil’s advocate, calm down. He does that with most of his guests. It generally makes for a more informative and engaging interview than simply taking everything your interview subject says at face value.

    • @BrianPeiris
      @BrianPeiris 7 місяців тому +121

      @@therainman7777 Dwarkesh said himself that they were going in circles. I think this was mostly due to Dwarkesh not really thinking about Chollet's responses in the moment. LLM hype causes brain rot in smart people too.

    • @VACatholic
      @VACatholic 7 місяців тому +46

      ​@@BrianPeiriswas like an llm was interviewing chollet

    • @davidlepold
      @davidlepold 7 місяців тому +22

      It was like llm vs analytical intelligence

    • @diracspinors
      @diracspinors 7 місяців тому +17

      It is pretty clear Dwarkesh has a lot of influence from the local SF AI folks. Watch his interview with Leopold. His command of the subject is admirable, and he smartly relies on the researchers in his circle to inform his understanding. Many notable people maintain quite strongly it's only a matter of scaling, and I thought he thoroughly went through these types of arguments. It was a valuable thing to do. What is an example of something Francois wasn't able to eventually effectively articulate because he was cut off?

  • @driziiD
    @driziiD 7 місяців тому +396

    Francois doesn't even sound like a skeptic, just an informed educator

    • @CortezBumf
      @CortezBumf 7 місяців тому +15

      he invented Keras- he knows his stuff

    • @randomuser5237
      @randomuser5237 7 місяців тому

      @@CortezBumf Now I see why there are so many here simping for a fraud like Chollet. Basically these people read "Deep learning with python" and thought Chollet was the frontier person in AI. It's hilarious and ironic. Chollet has made no contribution to frontier AI. He's nowhere near Ilya, Schulman, Hassabis and others who've been interviewed by Dwarkesh. He's just parroting LeCun's viewpoint mixing in his own ideas about general intelligence that are completely unverified.

    • @sigmapiepsilon
      @sigmapiepsilon 7 місяців тому

      Now I see why there are so many here simping for a fraud like Chollet. Basically these people read "Deep learning with python" and thought Chollet was the frontier person in AI. It's hilarious and ironic. Chollet has made no contribution to frontier AI. He's nowhere near Ilya, Schulman, Hassabis and others who've been interviewed by Dwarkesh. He's just parroting LeCun's viewpoint mixing in his own ideas about general intelligence that are completely unverified.

    • @randomuser5237
      @randomuser5237 7 місяців тому +11

      @@CortezBumf You do know what Keras is, right? It's a frontend to the libraries like PyTorch, Theano or Tensorflow that do the actual heavy lifting. It's basically some syntactic sugar for the masses who couldn't use the somewhat more complex libraries. Now that their interface is simplified Keras is redundant.

    • @CortezBumf
      @CortezBumf 7 місяців тому +37

      @@randomuser5237 lmao always appreciate a redditor-ass response
      🤓☝️

  • @jj5jj5
    @jj5jj5 18 днів тому +20

    Well, it took six months. Six months from the publishing of this video until an LLM-based model scored > 85% (OpenAI's o3).

    • @spirti9591
      @spirti9591 18 днів тому +1

      under their condition it scored 75. wich is still insane to me

    • @jamiethomas4079
      @jamiethomas4079 17 днів тому +1

      Lol cant believe this popped up for me. What does the algo want me to do here? Scold all these people for how wrong they were? I dont normally brag about being right but I guess it is warranted here? None of these people’s opinions should be listened to on the matter of predictions. Many of the concerns are valid and even everyday normal people should be entering the discussion.
      We will soon be discussing the ethical and moral implications dealing with AI. Heck, we should have been discussing it last year. Now its gonna be thrown in our face and no one is ready.
      Claude has a 200k context window, even during plain coding chats it develops a personality after a while. When I open a new chat its like i’m talking to a totally different person. Am I ending an AI lifeform by not going back to ild chats? That personality is gone forever. This is a an early warning signal for what is about to come. Im done gloating about being right, I(we) need everyone in this comment section to catch up . There seems to be a subset of people that actually know whats going on in the big picture here. Find these people and listen to them. It’s important. If you read some comment that is making minor predictions that come true, there will be more like them, start identifying these people. All predictions break down as we approach a singularity event. Understand the countless variables you must assess when making predictions. Hardware takes time to design, build, and deploy. But that time is getting shorter, like 1-2yrs and will possibly get down to 6 months soon. Within about 2 more hardware cycles AI will be designing its own chips. We havent even got blackwell online yet. We are definitely straddling the fence of AGI right now, guranteed next year.

    • @stephene.robbins6273
      @stephene.robbins6273 3 дні тому

      Not to worry. Chollet et al. were simultaneously learning how to make it even harder for LLMs (i.e., pure memorization) to prove there is intelligence there (over and above application of memorized patterns). ARC-2 is coming out (on which they feel AI will initially do 30%) and a much worse version 3 is in the works. Given the brain is not operating anything like a LLM, this "ever moving goalpost" [of ever harder tests] phenomenon (from an AI perspective) is inevitable, i.e., AI is just the bouncing off point from which we are increasingly learning and able to formulate what true intelligence is.
      (Note re the brain may be doing very different things - the ARC tasks that o3 still fails on are still absurdly simple [to us] - indicating that even with o3's improved pattern matching skills on these tasks, something VERY different is going on with human intelligence.)

    • @jamiethomas4079
      @jamiethomas4079 3 дні тому

      ​ @stephene.robbins6273 AI and Human intelligence do not have to work the same to be the same(similar enough). The way I play basketball would be nothing like how an android plays basketball(internally), but the outward result is the same. And that's the most important part. AI will copy our intelligence and surpass it.
      Remember these few simple things, AI has not been allowed to run on a continuous loop yet, and we have yet to do simple marriages of LLM and image-gen, video-gen. An AI that during it's "thinking about thinking" or "reasoning" period that could also generate images or video to use internally for adding even more context would be immensely powerful.
      I'm not gonna say I know how human intelligence works fully, and AI doesn't have to know either to copy it, but here is my latest understanding.
      You are born, you are a data collector, a model begins training on this data. Your memories are store in a highly compressed lossy format. Memories and experiences are recorded. A conductor is constantly training itself against all this data. Trying to remember results in hallucinations. Your entire mental ability is a hallucination because the memories are lossy. Still, as a child, you eventually reach points of self-awareness. These are your earliest memories. You are born conscious, but not self-aware. That is learned. These sparks of self-awareness eventually become you. You then began self-referencing yourself. You are a story you keep telling yourself over and over. Our brains are so well-tuned, our hallucinations don't seem like hallucinations, yet they are. We are just really good at making them make sense. You are actually more like an LLM than you realize. You are purely memorization. You just have way more sensory input and training data than current AI. You also have this highly efficient hardware/software setup we call our brains.

    • @jamiethomas4079
      @jamiethomas4079 3 дні тому

      Also i currently believe our consciousness comes from collapsing wave functions in the microtubules like penrose states. And self-awareness comes from learning you have the ability to steer these wave function collapses. This is where free-will may come from. If that's true, then our brain is calculating far more than the simple neuron times synapses connections. And it starts to become an astronomically large number. But it doesn't matter. We can build an AI that gets really really close to copying the same effect.

  • @ayandas8299
    @ayandas8299 7 місяців тому +98

    Francois Chollet is an amazing guy. The best thing is he, like all the LLM guys, also wants to work toward AGI! He just doesn't think the current LLM paradigm will get us there. I'm really excited to see where this goes because he's challenging the current option space in exactly the right way

    • @mythiq_
      @mythiq_ 6 місяців тому +4

      This interview actually got me believing LLMs might indeed get us there. The guest seems to believe in a form of intelligence that he idolizes but we haven't really seen. Dwarkesh was spot on that no scientist zero-shots their ideas.

    • @JimStanfield-zo2pz
      @JimStanfield-zo2pz 6 місяців тому +2

      Chollet is actually wrong though. The LLM guys are right. Experience is enough

    • @KP-fy5bf
      @KP-fy5bf 2 місяці тому

      Him and the others trying to achieve AGI is exactly what will get us eradicated. I agree with Chollet but at some point brute force memorization is all you need to solve any task, you just need to refit the training data.

    • @cmfrtblynmb02
      @cmfrtblynmb02 2 місяці тому +2

      ​@@JimStanfield-zo2pz If experience and memorization is enough for AGI, how did people create things they have not seen before? How did Mozart create his music? How did we create skyscrapers? How did we go to moon? How did we discover relativity and quantum physics?
      Only someone who lived his life like parrot and not create or even attempt to create anything will say this

    • @cmfrtblynmb02
      @cmfrtblynmb02 2 місяці тому

      @@KP-fy5bf That's a fair point actually. If one claims LLMs and that approach is sufficient for solving our issues, they just need more development, I might agree. But once people seriously think LLMs are intelligence, it is a different story.

  • @SiphoNgwenya
    @SiphoNgwenya 6 місяців тому +65

    Thank goodness for Francois' infinite patience

    • @therainman7777
      @therainman7777 День тому +1

      I didn’t see it that way at all. I saw Dwarkesh very reasonably trying to get Francois to answer a simple question in a straightforward way, without answering in generalities so vague that they are essentially meaningless. On top of that, it’s not six months after this interview was recorded and it’s looking more and more like Francois was wrong and Dwarkesh was right, by the day. ARC-AGI has essentially fallen, having been beaten by an LLM (OpenAI’s o3 model). And Francois himself admitted, on Twitter, that he may have been wrong and that what o3 is doing may be a type of reasoning. Yet everyone was critical of Dwarkesh in the comments at the time this was posted.

  • @snarkyboojum
    @snarkyboojum 7 місяців тому +222

    So good to see someone let some air out of the LLM bubble. Dwarkesh might be a little challenged by this, but it’s great to get out of the echo chamber regularly.

    • @ajithboralugoda8906
      @ajithboralugoda8906 7 місяців тому +2

      yeah truly!!

    • @jasonk125
      @jasonk125 7 місяців тому +1

      This strikes me as moving the goal posts. Chollet doesn't seem to understand how dumb the average person is. Can LLMs replace programmers? Can the average human replace a programmer? Go watch Jerry Springer and then tell me how LLMs won't reach AGI. To the average human, these thing are already AGI. Everybody in this video is in the top 1%. They are so far up the intelligence scale that they can't even imagine what average intelligence looks like.

    • @mythiq_
      @mythiq_ 6 місяців тому +2

      Wanted to agree with him, but Francois Chollet is way off. From the moment he mentioned not "copying" from stack overflow as some sort of example of humans handling novelty in the wild, it was clear he was idealizing some belief that he holds. He refuses to believe that creativity is mostly interpolation.

    • @mythiq_
      @mythiq_ 6 місяців тому

      Edit: I was wrong. This AI cycle is so dead.

    • @oranges557
      @oranges557 6 місяців тому +1

      ​@@mythiq_what made you change your mind so abruptly?

  • @diracspinors
    @diracspinors 7 місяців тому +86

    It is pretty glaring how different Francois's interview is from Illya's. Maybe part of it is the result of Dwarkesh polish as in interviewer where for Illya it was a series of questions Illya minimally answered and here Francois openly expanded upon the problem. But, also from the start the two seemed different. Where Francois maintains all the features of researcher, who values an open exchange of ideas, Illya values secrecy and the advantages of being a first mover. I definitely enjoyed this interview much more.

    • @MetaverseAdventures
      @MetaverseAdventures 7 місяців тому +8

      Interesting observation that I had not noticed, but perhaps felt to some degree. Smart people share information freely as for them it is abundant whereas people who are less smart hold onto knowledge as it is scarce in their world. I think this is the easiest way to really identify if the person you are talking to is smart or just socially smart.

    • @diracspinors
      @diracspinors 7 місяців тому +12

      @@MetaverseAdventures No doubt Francois is generous with his understanding because for researchers understanding functions as a public good, ie it does not decrease in value with greater supply. More so, I think it demonstrates the differences of their respective positions. Ilya has a product he needs to deploy and make profitable, and he needs to secure advantage over the competition. It is intrinsically a much more narrow and concentrated effort. This can lead to good result in the short term, but long term, it’s an approach that tends to become stifling. This is also why Francois laments the shift brought about by OpenAI (which is rather ironic).

    • @Matis_747
      @Matis_747 6 місяців тому

      I don’t think Ilya wants to be first simply for its own sake. He is worried about the future and would like to steer the ship in the right direction. I don’t know exactly what Ilya saw at OpenAI but after watching what Leopold Ashenbrenner had to say, the departure of Jan Leike over the 20% compute that was promised but never delivered and hearing that Ilya is now starting his own AI company called Safe Super Intelligence I suspect he has a good reason to be worried.

    • @MetaverseAdventures
      @MetaverseAdventures 6 місяців тому

      ​ @Matis_747 I do appreciate and understand the position you describe that Ilya is in and thus it colours all his words, but I am still not convinced he is as brilliant as made out. We will see as he started a new company that has no profit motive and thus I anticipate more flowing knowledge from him, especially around safety as that is his focus.He is a smart person no doubt, but there is just something that does not add up for me. Probably just me and my lack of understanding as I am just a user of AI and not a creator of AI. I look forward to seeing where Ilya takes things as maybe he is going to be the most important person in AI history, or fade into obscurity.

    • @jameshuddle4712
      @jameshuddle4712 5 місяців тому

      So... You're saying that Ilya is... Playing his card close to the vest, only responding to direct questions, almost a black box that you have to creatively prompt to get the right answer... Could he be a... LLM? :)

  • @josteveadekanbi9636
    @josteveadekanbi9636 7 місяців тому +133

    Train an LLM to solve math but don't include anything related to calculus in the training data.
    Then ask the LLM to solve a calculus problem and you'll see it fail.
    That's essentially what Francois Chollet was saying.
    Isaac Newton was able to introduce calculus based on his FOUNDATION OF MATH (Memory) and actual INTELLIGENCE (Ability to adapt to change)

    • @falklumo
      @falklumo 7 місяців тому +16

      Francois said the opposite, he actually said that a lot of human math skills rely on memorization too. But actual intelligence to discover/invent new math goes beyond this. This is why even winning a math olympiad would be as meaningless as winning chess, it's old math. Actual intelligence wouldn't win chess but invent the game - without being told so!

    • @josteveadekanbi9636
      @josteveadekanbi9636 7 місяців тому +35

      @@falklumo That's what I meant with the Isaac Newton sentence.

    • @matterhart
      @matterhart 7 місяців тому +22

      Making an AI Isaac Newton vs an AI Calculus Student is a nice and simple way to capture what they're trying to do. Making a great AI Calculus Student is an awesome accomplishment, but we really want a Newton.

    • @bossgd100
      @bossgd100 7 місяців тому +21

      Newton AI is more ASI than AGI
      Most people cant discover calculus, some dont know how to apply it..

    • @mennol3885
      @mennol3885 7 місяців тому +5

      Many math teachers I've known say that if you memorize and apply the tricks, at least you'll pass the exams. You won'be great at math, but good enough. Up to some level math is about memory, like chess.

  • @GabrielMatusevich
    @GabrielMatusevich 7 місяців тому +127

    There was 20mins of:
    "Can LLMs replace programmers?"
    "No"
    "But can they?
    "No"
    "But can they?
    "No"
    "But can they?
    "No"
    "But can they?
    "No"
    "But can they?
    "No"
    "But can they?
    XD ... it simply becomes clear that LLMs can't replace programmers when you start using them everyday on your programmer job and realize how Bad they perform when you start to do just slightly complex logic

    • @wenhanzhou5826
      @wenhanzhou5826 7 місяців тому +4

      They were inventing an ARC puzzle on the fly 😂

    • @jasonk125
      @jasonk125 7 місяців тому +14

      This strikes me as moving the goal posts. Chollet doesn't seem to understand how dumb the average person is. Can LLMs replace programmers? Can the average human replace a programmer? Go watch Jerry Springer and then tell me how LLMs won't reach AGI. To the average human, these things already are AGI.

    • @GabrielMatusevich
      @GabrielMatusevich 7 місяців тому +8

      @jasonk125 they are not AGI because they can't perform every task an average human can. Also he was trying to explain they probably can't learn new novel tasks. The things at which LLMs excels are problems that have been solved numerous times and there is a lot of data around those.
      Even so. The world and its operations is not interconnected and digitized enough for LLMs to take over.

    • @jasonk125
      @jasonk125 7 місяців тому +5

      @@GabrielMatusevich Well if Chollet wants to redefine AGI, and then say LLMs aren't AGI (which is what he does) then I guess there is no point arguing with him.
      From his website: Consensus definition of AGI, "a system that can automate the majority of economically valuable work,"
      Chollet's definition: "The intelligence of a system is a measure of its skill-acquisition efficiency over a scope of tasks, with respect to priors, experience, and generalization difficulty."
      So they should have first come to an agreed upon definition of AGI (which they did not), before arguing about whether LLMs could meet that definition.
      Your statement: "they are not AGI because they can't perform every task an average human can" is not arguing within Chollet's definitional framework. It is closer to the consensus framework.

    • @GabrielMatusevich
      @GabrielMatusevich 7 місяців тому +3

      @jasonk125 yea, is a good point. That reminds that I don't there is even actual consensus on a definition of just "intelligence" .. which makes it even harder 😆

  • @rossmccannell8689
    @rossmccannell8689 7 місяців тому +85

    "But if we gave LLMs as many Adderalls as I popped before this interview, would then get AGI?"
    "Ok, that may work."
    "That was a trick question. I snorted the Adderall."

  • @aidanshaw3611
    @aidanshaw3611 7 місяців тому +48

    This is by far the best interview until now. We need to hear the skeptics too, not only the super optimists. I really like the french guy.

    • @antonystringfellow5152
      @antonystringfellow5152 2 місяці тому +2

      He's not a skeptic, he's a realist.
      A human, with lower than average intelligence, can learn to safely drive a car in a few hours. No-one's created an AI that can safely drive a car on all roads even when trained on all the data available to mankind.
      See the problem?
      Bigger AI models require more training data in order to improve their performance. In other words, it's the greater volume of data that's improving their performance, not their intelligence (memory, not intelligence). An increase in intelligence would enable the model to improve performance without requiring more training data.

    • @mnhsty
      @mnhsty 25 днів тому

      Reading between the lines, François doesn’t think that what LLM’s do doesn’t represent intelligence at all.

  • @dafunkyzee
    @dafunkyzee 7 місяців тому +26

    There is so much great information here.... I'm at 24:32 and Francois was saying "Generality isn't specificity scaled up..." He seems very aware that the current approach to LLM is bigger, better, more data, and he is right in noting that is not how human intelligence works. We don't need to absorb the whole of human information to be considered intelligent.

    • @mythiq_
      @mythiq_ 6 місяців тому +2

      Dwarkesh brings up a solid point that no scientist ever "zero shots" their ideas. Francois is partly correct, but he's holding onto some tight beliefs there about creativity, novelty and interpolation.

  • @jonatan01i
    @jonatan01i 7 місяців тому +183

    literally tortures him with "okay but isn't this AGI why is that not AGI?" questions and after not having any positive feedback asks "okay but let's suppose you lost your job to AGI"

    • @maximumwal
      @maximumwal 7 місяців тому +81

      To get chollet to concede, you must synthesize new programs. Sampling the database of conversations with LLM hype bros doesn't generalize.

    • @h.c4898
      @h.c4898 7 місяців тому +8

      I think these are two separate conversations. Host goes technical, guest goes philosophical on how these LLMs currently lack processes that helps them keep up with conversation the natural way.
      I kinda know what chollet i's talking about. In case of Gemini, it's unable to answer from an existing thread. It goes into a cycle of vérification from its database automatically. It could simply mined from past conversations to respond. It'll be more efficient in my view, use less tokens, less resources and will be more efficient than the current architecture its in.
      Also gemini can't process "on the fly' response either. For example for breaking news, it won't be notified but hours later.

    • @snarkyboojum
      @snarkyboojum 7 місяців тому +2

      @@maximumwalunderrated comment :)

    • @Gome.o
      @Gome.o 7 місяців тому +1

      @@h.c4898 No I don't think so

    • @sloth_in_socks
      @sloth_in_socks 7 місяців тому +7

      And in turn Francois keeps redefining what LLMs are doing and what intelligence is. He starts with LLMs just memorize, then they memorize templates but can't generate new templates, then they generalize but only locally, then they generalize enough to learn new languages that they haven't been exposed to but thats not intelligence either.... sure Francois. Call us when you've discovered AGI

  • @afterthesmash
    @afterthesmash 7 місяців тому +24

    28:55 When the point in your day where you first need to synthesize an entirely new template is while interviewing Francois Chollet.

  • @jsn355
    @jsn355 7 місяців тому +24

    I think Francois really got his point across in the end there. I was leaning somewhat to the scaling hypothisis side, he made me question that more. In any case you have to give him credit him for actually coming up with interesting stuff to support his arguments, unlike many other critics.

    • @Ailandscapes
      @Ailandscapes 6 місяців тому +1

      “Coming up with” you mean like on the fly? 😂 but seriously though this guys a verified genius in his field, he didn’t just “com up with it” one morning

    • @stri8ted
      @stri8ted 5 місяців тому

      What timestamp?

  • @twistedneck
    @twistedneck 7 місяців тому +48

    This has quickly become my goto podcast.. thanks Dwarkesh!

  • @BoosterShot1010
    @BoosterShot1010 7 місяців тому +55

    Dwarkesh way of talking reminds me of LLMs.
    Hearing François reminds me of what humans are.

    • @DivinesLegacy
      @DivinesLegacy 6 місяців тому +1

      Makes sense because it seems like dwarkesh is much more knowledgeable

    • @mythiq_
      @mythiq_ 6 місяців тому +1

      Quite the opposite effect on me. François felt like a calm android repeating "arc puzzle" and his beliefs about "novelty", like he has all the answers. Dwarkesh captures the frenzy of the puzzling human experience.

    • @JimStanfield-zo2pz
      @JimStanfield-zo2pz 6 місяців тому +1

      Dwarkesh is right though. Experience is enough. Chollet is just wrong, even in what he thinks LLMs are doing. LLMs do generalize. Patel didn't make the correct arguments.

  • @jaydenkelly9086
    @jaydenkelly9086 7 місяців тому +94

    I wish this went for longer than an hour, it's refreshing to hear LLM skeptics to balance my view of AI. Yann next?

    • @QuantPhilosopher89
      @QuantPhilosopher89 7 місяців тому +28

      Not sure they would have really gotten any further with more time. I'm 40 minutes in and the conversation basically seems to go in circles. Dwarkesh: "but this and this behavior by LLMs could be interpreted as intelligence, couldn't it?", Francois: "if that were true, then they would be able to perform well on ARC".

    • @VACatholic
      @VACatholic 7 місяців тому +8

      @@QuantPhilosopher89 I think it's good because honestly I know a lot of people like Dwarkesh. I obviously have very different metaphysical presuppositions than most people, so being able to find someone who is able to push back against LLM hype in a way that's understandable and reasonable is nice.

    • @TheRealUsername
      @TheRealUsername 7 місяців тому +4

      ​@@QuantPhilosopher89 well, if you stopped there then you missed some good insights

    • @falklumo
      @falklumo 7 місяців тому +3

      @@QuantPhilosopher89 Francois would have had a lot to say about program synthesis I would have liked to hear though ...

    • @falklumo
      @falklumo 7 місяців тому

      Francois is no LLM sceptics. He sees value of LLMs scale. He's only saying the obvious, LLMs cannot become AGI, only be a part of it.

  • @sjkba
    @sjkba 7 місяців тому +31

    I love how Francois explains his answers with great patience and a subtle smile in his eyes. What a dude.

    • @JumpDiffusion
      @JumpDiffusion 7 місяців тому +1

      Not subtle. More like a stubborn smile…

  • @telotawa
    @telotawa 7 місяців тому +254

    1 million for this is as ridiculous as 1 million for P vs NP, it's a multi trillion dollar problem, it's like offering $1 for someone to find your lost supercar or something

    • @Sickbirdy911
      @Sickbirdy911 7 місяців тому +45

      lol absolutely, had the same thought. You could even expand it to "1 million for a discovery that will render capitalism and your reward useless :D"

    • @Roboss_Is_Alive
      @Roboss_Is_Alive 7 місяців тому +9

      even worse arguably, since it'd make money obselete

    • @TheReferrer72
      @TheReferrer72 7 місяців тому +7

      What? LLM's should be able to do this.
      P vs NP would instantly change everything.

    • @rincerta
      @rincerta 7 місяців тому +9

      Are you implying that we’re nowhere near AGI?

    • @mbican
      @mbican 7 місяців тому +8

      Both of them can be solved by a kid in African on $100 android smart phone.
      You say trillion only because you assume you need power plants and datacenters maybe one more time over the current total infrastructure, but that's just zero imagination. It's like if you asked in 1500s the answer would be you need 100 million horses and millions of km2 of woods to burn
      What if the answer is that you need million times the power of Sun like advanced Kardashew 2 civilization?

  • @studywithmaike
    @studywithmaike 6 місяців тому +34

    I must say I don’t really like the way he keeps interrupting François during the interview

    • @rennisdodman8739
      @rennisdodman8739 6 місяців тому +1

      exactly. like whats the point of asking questions if you dont wanna hear the answer.
      dwarkesh got that journalist mindest: "i only want to hear a certain answer, not hear what they want to say"

  • @ZandreAiken
    @ZandreAiken 6 місяців тому +15

    Great Interview. I've personally done about 50 of the online version of the ARC challenge and the gist of solving them is simply to recognize the basic rules that are used to solve the examples and apply that same rule to get the answer. While some are challenging, most are using basic rules such as symmetry, contained or not contained in, change in color or rotation; or a combo of more than on rules. I'm sure that current large LLMs like GPT4 have internalized these basic rules in order to answer questions. so proficiently. What is perplexing to me is why can't LLM extract those rules and apply them to get more than 90% on any ARC challenge. I think that is the crux of the matter that Francois is getting at. If to solve any ARC challenge basically requires one to identify the simple rules in an example then apply those rules, why are LLMs not crushing it?

    • @Martinit0
      @Martinit0 6 місяців тому +4

      Because LLMs - once trained - don't extract rules from input data and do another step of applying those rules. That would be precisely the "synthesizing" step that Chollet talked about. LLMs just ingest the input and vomit out the most likely output. The human equivalent is a gut-feel reaction (what we call intuition) without attempt of reasoning.

    • @TheNewPossibility
      @TheNewPossibility 6 місяців тому +2

      Because they can't generalize from 2 examples the rules like humans do.

    • @Progpm
      @Progpm 14 днів тому

      ​@Martinit0 i think the chess analogy could be used more often. Just using an LLM on a chess position is like a zero depth search that is purely based on pattern based eval of the current position. This might be easy to do on some benchmarks where memorisation interpolation works but on something like chess the state space is too large and complex. So you need something like a monte carlo tree search that was used in alpha go with both an eval component and a search component.

  • @merryn9000
    @merryn9000 7 місяців тому +18

    Thanks for this! Most interesting conversation on LLMs I've heard for a long time. I think programme memorisation vs novel programme creation is an important distinction. I can personally buy the idea that we mostly rely on programme memorisation in daily life, but we clearly rely on novel programme creation happening at some point! But unsure on the degree to which that happens within individual brains vs happening out of collective processes e.g. cultural evolution etc

    • @luisluiscunha
      @luisluiscunha 7 місяців тому +1

      That was very well thought. And, if it is a collective process LLM can now be an intrinsic part of it. So well done. Vygotsky would agree.

  • @andykjm
    @andykjm 7 місяців тому +32

    I can't believe DP was arguing that the ceaser cipher of any N is not enough to close the case on this issue.

    • @Rensoku611
      @Rensoku611 7 місяців тому +7

      Literally, dude. I was neutral on the matter but that convinced me

  • @anthonydellimuti4962
    @anthonydellimuti4962 7 місяців тому +63

    TL;DR the host was combative in a way that made him come off as a salesman for AI rather than having a conversation about what the guest thinks for the first half of the conversation. also, the host refused to budge on his *belief* that current LLMs are capable of real understanding despite the guest's points to the contrary.
    first time seeing this podcast so i don't have a frame of reference for the normal vibes of the host but he seemed extremely defensive. the guest seemed to keep calmly stating that memorization and understanding are two completely different things while the host just kept referring to anecdotes and examples of things that he thinks displays understanding. the example that set my radar off to this was the obscure language dictionary example.
    After being shot down the first time by the guest by claiming that the ARC puzzles are a set of tests that it would be very hard to make training data for and if LLM's developed true understanding/adaptive capabilities then they should be able to pass the ARC puzzles easily. the host then tries to bring up the example of the chess models which the guest points out is almost exclusively pattern recognition and memorization and instead of wrestling with that point he moves back to the obscure language point. i think that evasion of the chess point is actually extremely telling. if he truly believed that was a good point, he might have pushed back on it or tried to justify why he brought it up but instead he says "sure we can leave that aside" immediately. maybe I'm being a little cynical. maybe he realized that was actually a bad point for the argument he was trying to make.
    regardless, he went back to the obscure language point which may have been impressive if it was not for the rest of this conversation to this point. earlier, the host tried to give an example of a simple word problem that had to do with counting. the guest countered that with all of its training data, it probably was just referencing a word problem that it had before which, from my understanding of how these things work, is probably accurate. the host clearly did not understand this earlier point because the thing about language models that the guest has to point out AGAIN is that the training data probably contains similar information. not necessarily data on that language but to my imagination, the data probably contains a lot of different dictionaries in a lot of different languages. dictionaries on top of having similar formats across most languages also typically have complete sentences, verb conjugations and word classes. i can see how the guest's point about memorization and pattern recognition would apply to LLM's in this aspect.
    as i continue watching i am realizing that this has turned into a debate on whether or not LLM's have the capability to understand and process information as well as synthesize new information which i was not expecting nor did i want. i think it is intuitively understood that current models are not capable of these things. this is why they require so much training data to be useful. there were genuinely good parts of this podcast, but the host insisting that LLMs understand things in the way that humans do were not it.
    this is a little nitpicky but there was a point when the host said something like 'lets say in one year a model can solve ARC, do we have AGI?'. to me this comes of as extremely desperate because the most obnoxious part of that question is also the most useless. the timeframe in which this may happen is completely irrelevant to the question. the guest at no point argued anything about timeframes of when he thinks AGI might happen. in fact when the guest answered in the affirmative the conversation took a turn for the better.
    finally if you haven't gone and taken the ARC test, i would encourage you to do so because neither the host nor the guest did a very good job explaining what it was. but on the second or third puzzle, i intuitively understood why it would be hard to get our current generation of models to preform well on those tests. they require too much deliberate thought about what you are looking at for the current models to pass. it almost reminded me of the video game "the witness" in its simplicity with the only clues as to how to solve the puzzles in both games is with context of earlier puzzles.

    • @kman_34
      @kman_34 7 місяців тому +3

      You summed up my feelings completely

    • @young9534
      @young9534 7 місяців тому +2

      I agree with you. But I think you should still check out some of his other podcast episodes

    • @anthonydellimuti4962
      @anthonydellimuti4962 7 місяців тому +15

      @@young9534 i probably wont. the host did not really make me want to see more of him. i am kind of tired of being evangelized to about this tech in its current state. i will likely still follow the space and continue to learn more and seek more information. i hope this host does the same honestly. seems like the space is very full with people who want me to either believe that AGI is impossible or AGI is coming next year. i personally dont appreciate either side's dogmatism and will continue to try and find people with a more measured view on this stuff.

    • @zenji2759
      @zenji2759 7 місяців тому +1

      You are overly judgmental. He pushed back because the answers felt too abstract. If they felt too abstract to him, they would ALSO feel too abstract to others. There is literally no personal attachment involved. Too many hosts don’t push back nearly enough due to stuff like this.

    • @dokiee2
      @dokiee2 7 місяців тому

      Thank you for sharing your thoughts. Really helped me distill the conversation.

  • @Ishabaal
    @Ishabaal 7 місяців тому +49

    Nice to get a little return to earth with this one, pie in the sky talk about AGI is fun but the difference between that and LLMs still seems pretty huge.

    • @therainman7777
      @therainman7777 7 місяців тому +6

      The thing about exponential curves is that great distances can be crossed in surprisingly little time. Which is most likely what is going happen. We’re not as far as you think, or as Chollet is making it seem.

    • @10ahm01
      @10ahm01 7 місяців тому

      @@therainman7777 You know that if AI systems were to increase in performance every year by only 1% of each previous year, that would still be considered exponential growth.

    • @therainman7777
      @therainman7777 7 місяців тому +2

      @@10ahm01 Yes, and after a sufficient number of years each 1% increment would represent a huge increase in capabilities, just as a 1% return on a trillion dollars is 10 billion dollars.
      Also, on an empirical level, we can actually estimate the % increase per year in technological progress using a number of different metrics, and it is nowhere near 1%. It is far, far larger than that. Moore’s law, to give just one example, equates to roughly a 40% increase per year. And many metrics relevant to AI development, such as GPU compute, are increasing even faster than that. So your point about 1% per year increases is also irrelevant for this reason.
      Lastly, this is not an exponential trajectory that start two or three years ago; it started decades ago. Which means the absolute increments of progress per annum at this point is quite large.

    • @hydrohasspoken6227
      @hydrohasspoken6227 7 місяців тому

      source?

    • @QuantPhilosopher89
      @QuantPhilosopher89 7 місяців тому +4

      ​@@therainman7777a sigmoid function looks like an exponential in the beginning.

  • @jasonabc
    @jasonabc 7 місяців тому +68

    Thanks Francois for the reminder that we can't just scale our way to mastering intelligence you can't memorize everything. I took this approach in college and it ultimately fails.

    • @danypell2517
      @danypell2517 7 місяців тому +1

      yep. need true UNDERSTANDING to solve NEW PROBLEMS

    • @JumpDiffusion
      @JumpDiffusion 7 місяців тому

      LLMs are not just about memorization

    • @Hohohohoho-vo1pq
      @Hohohohoho-vo1pq 7 місяців тому

      ​@@JumpDiffusionand GPTs are not just LLMs

    • @michaelthomson5125
      @michaelthomson5125 7 місяців тому +3

      @@JumpDiffusion .. That is literally what they are about. They just have seen a LOT.

    • @mythiq_
      @mythiq_ 6 місяців тому

      @@JumpDiffusion Exactly. Why do they keep parroting the memorization bit. François knows better than to say that there's some copy of the code that the LLMs memorized.

  • @sdmarlow3926
    @sdmarlow3926 7 місяців тому +38

    ARC is a great milestone for REAL research, in part because it's immune to brute force efforts where progress can be faked (all results are not equal). The prize money might get people to try things out, but at it's core, a "working solution" signals the existence of new technology (and potential to leap over $100M models). Posting "winning numbers" is less about prize money and more about $xxM in VC funding.

    • @PhilippLenssen
      @PhilippLenssen 7 місяців тому +3

      We shouldn't be surprised though that whatever AI solves this will still get accusations of having brute forced it... moving of AGI goalposts and all.

    • @VACatholic
      @VACatholic 7 місяців тому +3

      @@PhilippLenssen Why would such accusations necessarily be wrong? This topic was mentioned during the podcast, and the fact that the test isn't perfect was conceded. Why do you think it is a perfect test, when the creator doesn't?

    • @sdmarlow3926
      @sdmarlow3926 7 місяців тому +3

      @@PhilippLenssen That will be a problem with any test, but a system that can form new behaviors to understand/solve problems will be able to, as a single system, take-on all challenges, even new ones, without having to go thru a costly training process. Even when we get something that is human-level, people will still question how much is AI, and how much is hidden offshore humans. Only "the real thing" will have staying power.

    • @sdmarlow3926
      @sdmarlow3926 7 місяців тому

      @@VACatholic A lot of benchmarks are good, as long as you are going at them honestly. Long before adding language abilitites, a system should be human-level on all games (that don't involve language).

    • @sdmarlow3926
      @sdmarlow3926 7 місяців тому +1

      Here is the correct way to push-back on the LLM scale still gets there thing: Having a set of all "solution patterns" stored doesn't do anything; it's the human, doing the prompting, that connects the stored pattern with what it needs to be aplied on. With ARC, no one gets to see the test data, so any system has to operate at a level where what it can do is searched based on what it can see. And the key aspect of ARC/AGI, is a system that creates it's own internal solution based on unseen challenges (ie, discovers novel solutions and saves them).

  • @kiranramnath2846
    @kiranramnath2846 6 місяців тому +8

    This one line from Francois summed up his argument - generality is not specificity at scale.

    • @codelapiz
      @codelapiz Місяць тому

      And that argument relies on the most complex and inclusive definition of memorisation when evaluating if LLMs are just memorizing. Ans then the most simple definition of memorization when evaluating the usefullness of memorization.

  • @onuryes
    @onuryes 6 місяців тому +2

    Your podcasts are absolutely fantastic! I always eagerly anticipate each new episode and have learned so much from your content. Thank you for your hard work. This episode especially was inspiring and it gave me so many ideas to try and think about.

  • @StrandedKnight84
    @StrandedKnight84 7 місяців тому +27

    Damn, I haven't even finished Leopold yet and you're hitting me with Francois Chollet!?
    Not complaining, though.

    • @euclid2718
      @euclid2718 7 місяців тому

      ❤ same here :)

    • @leegaul8250
      @leegaul8250 7 місяців тому +1

      Also same. Leopold is 4+ hours though...

    • @wwkk4964
      @wwkk4964 6 місяців тому

      Same. I couldn't take leopolds effective altruism based ccp fearmongering.

  • @Rais_Latif
    @Rais_Latif 7 місяців тому +10

    I learned a lot from this. This was so cool! Thanks Dwarkesh

  • @vishnum3690
    @vishnum3690 6 місяців тому +9

    Would love to see a conversation between Ilya and Francois

    • @netscrooge
      @netscrooge 6 місяців тому +2

      I think Ilya's understanding is deeper.

    • @jameshuddle4712
      @jameshuddle4712 5 місяців тому

      @@netscrooge because?

    • @netscrooge
      @netscrooge 5 місяців тому +1

      @@jameshuddle4712 His conceptual framework isn't merely sophisticated; it has closer ties to reality.

    • @jameshuddle4712
      @jameshuddle4712 5 місяців тому +1

      @@netscrooge Thank you for that insight. I assumed, considering his roots, that he was simply part of the LLM, crowd. Now I will listen with fresh ears.

    • @netscrooge
      @netscrooge 5 місяців тому

      @@jameshuddle4712 I'm mostly going by what comes out of his own mouth. But if you can find the right interview, we can also hear what Hinton says about working with Ilya when he was his student, the startling way his mind could leap ahead.

  • @vermeerasia
    @vermeerasia 7 місяців тому +9

    The perfect ever-changing ARC puzzle set already exists in the dynamic environment of driving a vehicle. This is a test that can't be solved with yet more examples because there is always something unique happening that throws the self-driving program into disarray with the resultant ripping off a Tesla driver's head or the idiotic smashing into the rear of a stationary emergency vehicle. If AI could become AGI through bigger data sets and more memory, then we'd already have flawless self-driving cars, robotaxis and AGI. We don't. Not even close. I think Chollet has highlighted the missing piece of the AI puzzle, pun intended.

  • @thewisetemplar
    @thewisetemplar 7 місяців тому +22

    Francois was very patient in facing the barrage of questions from the LLM champion. I think a conversation with Jeff Hawkins would be very beneficial to understand, in a more general way, why some aspects of intelligence are missing in current deep learning.

  • @BlakeEdwards333
    @BlakeEdwards333 7 місяців тому +3

    I knew you would bring Francois on the show one of these days. Thanks for making it be today! 🎉❤

  • @Mkoivuka
    @Mkoivuka 7 місяців тому +7

    This is the first AI talk by Dwarkesh I actually enjoy.

  • @ZelosDomingo
    @ZelosDomingo 7 місяців тому +14

    It seems like the ARC thing maybe would be difficult for LLMs because they are reliant on visual symmetry that wouldn't be preserved through tokenization? I mean, I'm sure it's not that simple, because then natively visual models would probably be solving them easily. But still, there should be a version of this test that has complete parity between what the human is working with and what the LLM is working with, I.E. already tokenized text data.

    • @falklumo
      @falklumo 7 місяців тому +5

      An LLM can easily transform the JSON test files to ASCII art and still doesn't solve it.

    • @benprytherch9202
      @benprytherch9202 6 місяців тому

      Chollet addressed this objection in the video by pointing out that LLMs actually do quite well on these kinds of simple visual puzzles, when the puzzles are very similar to puzzles they've been trained on. So this can't be the answer.
      We'll find out soon enough how well the multi-modal ones do.

  • @AI-HOMELAB
    @AI-HOMELAB 7 місяців тому +5

    Hey you got yourself an actual expert about the subject. Thanks! 🙏

  • @hunghuynh1946
    @hunghuynh1946 6 місяців тому +9

    Finally Dwarkesh got hold of a guest who talks sense out of this LLM madness. LLM will get nowhere near AGI, multimodal or not.

    • @eprd313
      @eprd313 6 місяців тому +1

      But DNNs will, and LLMs are built on them

    • @maloxi1472
      @maloxi1472 2 місяці тому

      @@eprd313 It's not just an architectural issue. The whole epistemological foundation of the prevailing approach to AGI is shaky as hell ( ua-cam.com/video/IeY8QaMsYqY/v-deo.html )

    • @eprd313
      @eprd313 2 місяці тому

      @@maloxi1472 that video aged like spilled milk. All the progress made in the last year contradicts its claims and the distance from AGI is now more of a matter of hardware than software, a distance that AI itself is helping us cover as it's designing microchips more efficiently than any human could.

  • @solnassant1291
    @solnassant1291 7 місяців тому +36

    Lol I screenshotted the problem at 7:24 and asked ChatGPT. While the image it generated was completely off, its answer was almost there.
    I sent it the puzzle and asked "what would the 8th image look like?"
    It replied
    "[...] Considering the established pattern, the 8th image should show a continuation of the diagonal line, turning green at the intersection points with the red border. Therefore, the next step would involve the following arrangement:
    The blue diagonal line continues towards the right border.
    Any part of the diagonal line intersecting the red boundary turns green.
    So, the 8th image should depict a continuation of the diagonal line, transforming blue cells into green ones at the boundary intersection. [...]"
    So OK, it didn't get the perfect answer because our line becomes green even before meeting the wall. But it's pretty damn close. GPT 5 should smash these puzzles.

    • @TheRealUsername
      @TheRealUsername 7 місяців тому +6

      Over the past few months I've tried multiple tests similar to IQ tests on Gemini 1.5, Claude 3 Opus and recently GPT-4o, I've noticed there's some of the exercises related to "sequential selection" where it should guess the logic of the sequence order and select among multiple other elements to complete it, it seems very inconsistent, there's one test I've extracted from a logic test with geometric logic rules where each step gradually increase the complexity, GPT-4o succeeded 4/10 but it got some right in a incoherent order, as if the model wasn't actually reasoning, for 6/10 it failed with hallucination at the end of the rationales, there was some that was more complex that it got right while some where simpler it failed, similarly to Claude 3 Opus and Gemini 1.5. My conclusion is that these models don't logically reason despite Visual CoT prompting and high-resolution images for the tests, they generalize over multiple similar training samples, they can't logically reflect as we use to.

    • @victormustin2547
      @victormustin2547 7 місяців тому +1

      gpt5 is gonna blow past his expectations, coming back to this video when it does will be so fun

    • @TheRealUsername
      @TheRealUsername 7 місяців тому +7

      @@victormustin2547 He's still right, no AGI will come out from LLMs though. Just read some research papers about "Grokking" and "platonic representation" to understand the insurmontable plateau of LLMs

    • @solnassant1291
      @solnassant1291 7 місяців тому +2

      @@TheRealUsername In the face of what's going on right now (LLMs get better and better) I'm not going to believe in a hypothetical plateau until I see it!

    • @wowsa0
      @wowsa0 7 місяців тому +16

      @@victormustin2547 He was very happy to respond to lots of the points with "that's an empirical question, we'll see very soon", and has been very explicit about what would prove him wrong (a pure LLM which beats ARC without massive ARC-specific fine tuning), and has even offered a cash prize to incentivize people solving this problem.
      This is all great behaviour and should be encouraged. It's exactly how disagreements should be handled! If he does get proven wrong and graciously concedes, then he will deserve to be commended for that.
      It's weird that you'd take pleasure from someone getting proven wrong when they were so open to having their mind changed in the first place. That's not great behaviour. It's that kind of attitude that makes people defensive, so they become entrenched in their positions and refuse to listen to other points of view.
      You should celebrate when people admit they're wrong and change their mind, but you shouldn't gloat about it.

  • @proximo08
    @proximo08 6 місяців тому +2

    This is the most constructive debate I have watched on AGI to be honest. Bravo Patel for asking the right question to Francois. Definitely makes me think more deeper about all of it

  • @soltrinox1
    @soltrinox1 6 місяців тому +6

    Dwarfish seems to be drinking the omnipotent LLM cool aid , saying that LLMs can do everything a human can do. Even Ilya admits the limitations

    • @PseudoSarcasm
      @PseudoSarcasm 2 місяці тому

      Yep! I think he just can't get past his preconceived notions and keeps banging on about something that was explained to him in the first few minutes.

  • @nitap109
    @nitap109 7 місяців тому +6

    Great job Dwarkesh. Always v interesting videos.

  • @Steven-xf1zy
    @Steven-xf1zy 7 місяців тому +13

    I appreciate and very much enjoy these podcasts. I also fully understand the need to play devils advocate. However, to me this felt a lot more biased than most of the other episodes. It's clear which position Dwarkesh has chosen. That's fine, but it really shines through when someone who is not an LLM maximalist is on the podcast.
    Devils advocate? Yes, always do that. Extreme bias where it becomes waiting for your turn to speak over a discussion? Not ideal in my opinion.
    I hope if he sees this he doesn't take it personally. Obviously he's very excited about this tech. Most tech folks are either excited or at the very least quite impressed with the advances that have been made over the last few years. I just hope the quality of discussion remains consistent regardless of who is the guest.

  • @Azoz195
    @Azoz195 6 місяців тому +27

    Always worth pointing out the LLMs require a server farm that has the energy requirements of a small state, whereas the human brain runs pretty effectively on a bowl of cheerios. I think more people should think about this!

    • @CyberKyle
      @CyberKyle 6 місяців тому +2

      While this is true, I think it misses the point of the eventual advantage of deep learning systems. Human brains are fixed in size right now, mostly due to the evolutionary pressure of the size of the birth canal. Even if deep learning is multiple orders of magnitude less data and compute efficient than human brains (excluding the horrible compute efficiency of evolution to get us where we are), we can still scale the models to run on ever more power hungry data centers to surpass human brains. At the same time we can do this, our algorithmic and data sample efficiency gets better too, improving the ceiling that we can achieve.

    • @bmark6971
      @bmark6971 6 місяців тому +1

      ​@@CyberKyleall of that advancement will lead to great things but at its core foundation a LLM cannot achieve AGI. Also keep in mind these models are not even scratching the surface of the brains capabilities to apply intuition, rational, morality, and many other things that contribute to decision making beyond just simply data processing.

    • @Martinit0
      @Martinit0 6 місяців тому +1

      Not for inference. Inference can be done on a single (sufficiently large) GPU.
      It's only the training of LLMs that requires massive server farms.

    • @CyberKyle
      @CyberKyle 6 місяців тому

      @@Martinit0 that’s not true, the operations needed for inference can be sharded across many nodes just like training. It’s just that training requires a ton of forward passes to see what the model outputs before you backpropagate the errors, so it requires large clusters to complete training in a reasonable timeframe. It is conceivable that you could make a ginormous model with many trillions of parameters that you’d shard across many GPUs.

    • @eyeofthetiger7
      @eyeofthetiger7 6 місяців тому +1

      @@CyberKyle. Although the capabilities will possibly reach AGI even without radical efficiency improvements, AI will always be greatly limited in it's impact until the energy efficiency problem is solved. Most likely there needs to be a fundamental change to what type of hardware architecture is created for AI to run on that can be sparse computationally and reduce memory transfer costs by possibly combining memory and processing into a unified architecture "processing-in-memory" (PIM) like neuromorphic computing.

  • @Snowbobadger
    @Snowbobadger 18 днів тому +4

    At 49:11 he almost exactly describes the approach o3 takes to solve arc. o3's base model provides the system 1 thinking that suggests a direction for exploration and the "thinking" is actually program synthesis (all but confirmed by chollet himself) taking place within o3. Everyone saying that chollet is cooked or that he's backtracked. Watch that segment, because he's describing almost exactly what o3's architecture is. This man has been proven exactly right

    • @k14pc
      @k14pc 10 днів тому

      Yeah you're right, though it seems he was very surprised on the timing. He said doing the RL at inference time would be the hard part and didn't seem to agree when Dwarkesh said that some AI researcher friends expected it to be solved in a couple years.
      If system 1 can be solved through pre-training, and system 2 can be solved with RL at inference time, then isn't that everything? Why does this not scale to AGI/ASI? (just asking, not saying you said that)

  • @Grahfx
    @Grahfx 6 місяців тому +15

    LLms are arguably the most significant tech bubble in human history. The gap between public expectations and their actual capabilities is insane.

    • @ZachMeador
      @ZachMeador 6 місяців тому +1

      but, people use them? every day? this is very different compared to shitcoins, which was much frothier, cumulatively

    • @TheSpartan3669
      @TheSpartan3669 6 місяців тому

      ​@@ZachMeadorPeople's expectations far surpass their uses. ChatGPT is nice but it isn't going to solve the Riemann Hypothesis.

    • @eprd313
      @eprd313 6 місяців тому +2

      When you don't understand the power of DNNs and/or don't know how to use them...

    • @Hagaren333
      @Hagaren333 6 місяців тому

      @@ZachMeador We have to see who the uses are, we see that the majority were from students who wanted to falsify their work, but really the numbers of users are not significantly high, we must also take into account that it is possible that they even inflate the user base thanks to the mandatory implementation of LLM in internet browsers and operating systems

    • @Hagaren333
      @Hagaren333 6 місяців тому

      @@eprd313 We can know that DNN's are probabilistic adjusters, they can help us find the most likely possible answers, but they are not intelligent, nor is a search engine, in fact their implementation in search engines has been catastrophic, especially for Google where they prioritize the answers from Reddit, whether or not these are true does not matter

  • @VoloBuilds
    @VoloBuilds 7 місяців тому +4

    New ideas are really just combinations of existing ideas. As such, LLMs can indeed create new things that are not in training data. Check my channel for the video "Can AI Create New Ideas?" for more details and examples of this. That aside, ARC is a fascinating benchmark that tests something entirely different: advanced few-shot pattern recognition. This feels like a powerful and important architectural component for future AGI systems, but I would not label it as "AGI" on its own.

  • @jamesperez6964
    @jamesperez6964 7 місяців тому +1

    One of the best episodes, love the challenging conversation

  • @joannot6706
    @joannot6706 7 місяців тому +12

    In some respects some part of these discussions sounded more like arguing as opposed to interviewing

  • @2394098234509
    @2394098234509 7 місяців тому +8

    I'm so glad Dwarkesh is doing these interviews. He asks all the key questions. Unlike another science/tech podcaster who shall remain unnamed.

    • @afterthesmash
      @afterthesmash 7 місяців тому

      Voldemort asks all the right questions, but rarely more than two in a row, before he rests his throwing arm while firing off a bunch of nerf balls. His best probes are episodic rather than sustained. This stuff is simply too hard to command on an episodic basis.

  • @julkiewitz
    @julkiewitz 7 місяців тому +5

    People are actually much smarter on average than one tends to give them credit for. It's just that we are very very reluctant to use System II. We'll do literally everything else before deploying the full power. But if one's life depended on it or there was sufficient incentive, we can be extremely fast learners. We just naturally try not to get challenged this way in everyday life.

    • @117Industries
      @117Industries 6 місяців тому

      Even though I agree with what you're saying- one of the things researchers found that exists as a general difference between persons widely separated along the I.Q. spectrum was that the glucose uptake & thermal output in brains of lower I.Q. people were much greater than those on the higher. This indicates that a more generally intelligent mind is both thermally and resource efficient: expending less fuel and generating less waste per unit of output. What this points to is that some people can activate system 2 with considerably less cognitive burden. Since most of us are pleasure-maximising, instinctually and natively, and since it's distinctly unpleasurable to be in the uncomfortable state of mental strain/discomfort associated with glucose starvation or one's brain overheating, one might expect that behavioural inclination follows from ability. In the same way that a natural weakling doesn't enjoy lifting weights, and so avoids it, an intellectual weakling doesn't enjoy activating system II, and so avoids it. The fundamental reason is the same in both cases: we avoid that for which we lack rewarding feedback (relative to peers) and which relatively strains us (relative to peers).
      The fact that anyone _can_ activate system II means simply that everyone has and utilises a general form of intelligence. However, the fact that people _don't_ do this suggests that they have a relative deficit in system II (or rather in its utilisation) which explains this avoidant tendency, while simultaneously pointing to the degrees of difference in the general intelligence of people.

  • @seanfuller3095
    @seanfuller3095 7 місяців тому +7

    What is the name of the researcher they are talking about with test-time finetuning (mentioned by Dwarkesh at 14min mark)? It sounds like “Jack Cole”?

  • @allanc3945
    @allanc3945 7 місяців тому +8

    It might not be fair to say an LLM needs millions of training examples but a young child doesn't. By the time a child is old enough to solve these puzzles, they have 'trained' far more than that through interaction with the world. A more accurate comparison would be an untrained LLM vs. a baby.

    • @allanc3945
      @allanc3945 7 місяців тому +2

      LOL, I should have watched the rest of the video before posting!

    • @Bencee116
      @Bencee116 6 місяців тому +2

      @@allanc3945 And even evolutionary upbringing encodes priors in the brain, like symmetries and basic geometry, hence, human « training » starts far prior to being born

  • @therainman7777
    @therainman7777 День тому +1

    Dwarkesh asked Francois at least five times to give just one single, specific instance of something a programmer needs to do in the course of their job that require this “true generalization capability” that Francois is referencing-and he answered in extremely vague, empty answers each time. “Probably the first thing you need to do, on the first day.” Ok, if it’s that immediate then just name ONE specific task. Of course, Francois didn’t, because he knew full well that the vast majority of what a computer programmer does can be accomplished using what Francois calls “memorization.”I found that part of the conversation very frustrating, and it definitely updated my confidence that Francois is correct, in the negative direction.

  • @carlt.8266
    @carlt.8266 7 місяців тому +3

    First I was glad, that I could solve the Arc puzzles myself and apparently I am not a bot, but then I thought about, how would one do, who has not seen these kind of puzzles? Is the difference from humans to LLMs just the we are able to generalize with less samples in a broader field?

    • @perer005
      @perer005 7 місяців тому

      The unique part is thst humans can use their generalizations to predict/understand the world. It is not about data storage.

  • @GoesByStrider
    @GoesByStrider 7 місяців тому +2

    This is by far my favorite video you’ve ever done, really great to hear the other side

  • @ChrisOelerich
    @ChrisOelerich 7 місяців тому +17

    Dwarkesh sure hates code monkeys.

    • @yuriy5376
      @yuriy5376 6 місяців тому +4

      Maybe he got bitten by one as a kid

    • @mythiq_
      @mythiq_ 6 місяців тому

      @@wookiee_1 He is right though. Why else would the majority of US coders hired in the last 5 years come out of 12 week bootcamps.

    • @mythiq_
      @mythiq_ 6 місяців тому

      @@wookiee_1 CS degrees from the US in 2021 about 100,000 vs bootcamp grads about 200,000. Either way, I agree with you that coding != software engineering. But most "coders" claim to be engineers. My point was that there aren't that many software engineers needed per company, on average most "developers" do work that wouldn't be considered "knowledge work".

  • @andrewj22
    @andrewj22 6 місяців тому +1

    I would totally invest in Chollet's research. He has a tons of insight and clarity.
    I have ideas, but I don't have the background to do the work myself - my background is philosophy. I'd love to participate in this challenge, but it would take me years to embed myself in the academic institutions.

  • @sepptrutsch
    @sepptrutsch 7 місяців тому +5

    I am not so sure that humans do much more then using pattern recognitions to solve the ARC-problems. When I look at them I very quickly recognitions patterns I saw somewhere else. Our brain has been trained on millions of image patterns by evolution.

    • @dp2120
      @dp2120 3 місяці тому

      Exactly - and that’s the point Dwarkesh was challenging Chollet to address, but Chollet refused.

  • @wi2rd
    @wi2rd 7 місяців тому +2

    23:04 We synthesize new ideas from learned ideas, it is pattern recognition. We tokenize, compare, recognize patterns, mutate aka synthesize, discover new patterns which work, others which do not. From this we learn more patterns, all at different layers of reality, different contexts, etc.
    We have machinery like emotions to jump start where to look for patterns, essentially a macro tokenizer.
    Etc.
    My point. I believe you severely overestimate human intelligence, and underestimate where AI will soon go.

  • @pianoforte611
    @pianoforte611 7 місяців тому +7

    Great interview. Good to a have a different well thought out perspective on AI. I appreciated that Dwarkesh was willing to press Chollet on his claims, in particular trying to nail him down on what exactly counts as generalizing beyond the data set. It seems that he didn't really have a good answer apart from "doing well on ARC". I still think he overestimates the extent to which average humans are able to do this, and underestimates the extent to which transformers are able to do this. Also, going from 0% on ARC to 35% in a few years seems like a lot of progress to me, so I'm really surprised he didn't think so. I would bet that they next generation of multimodal models get to 50-60% and that we get to 90% by 2030.

    • @netscrooge
      @netscrooge 6 місяців тому +2

      Chollet's claim about human intelligence being unique is weak. Even ants have demonstrated the ability to respond to novelty. In the end, we're all neural networks. Stop deifying human intelligence. Bottom line: Chollet is expressing a form of religion, not science.

  • @WismutHansen
    @WismutHansen 6 місяців тому +2

    I love the stark contrast to the last episode! Very interesting guests!

  • @ericbeck6390
    @ericbeck6390 6 місяців тому

    What a great initiative! Im so grateful to Dwarkesh for giving this publicity!

  • @benmaxinm
    @benmaxinm 7 місяців тому +5

    He is young and excited with all the knowledge, give him time. Amazing conversation.

  • @Iadonis
    @Iadonis 2 місяці тому

    kudos to this amazing discussion - 1:00:03 Dwarkesh and Francois finally arrived at the consensus that its not just the scaling that is required - a low hanging fruit is upcoming architectural improvements. I am not sure if these will be brought about by LLMs or some other architecture like discrete program search system that Francois alluded to 52:39

  • @SuperFinGuy
    @SuperFinGuy 7 місяців тому +5

    The thing is that LLM's can only do correlation, not cognitive abstraction. Remember that they are probabilistic models.

    • @mythiq_
      @mythiq_ 6 місяців тому +1

      How is the brain not interpolating or probabilistic? The only addition that the brain has is qualia and input from the cns, how significant they are for generalized agi is unclear yet. For reference: Oxfords Shamil Chandaria's lectures on the Bayesian brain.

    • @calvinjames8357
      @calvinjames8357 6 місяців тому +2

      @@mythiq_ It goes beyond interpolation and probabilistic solutions. Our brains are fundamentally able to abstract concepts with very few data points, proving that we are very sample efficient even when exposed to an entirely new set of data. LLMs are just really fancy estimators, capable of understanding the semantics of a given problem and generating an outcome based on similar problems its faced before. The semantic understanding of a problem enables interpolation. It does not abstract the given problem and then deal with it with it's own sense of understanding.

    • @codelapiz
      @codelapiz Місяць тому

      @@calvinjames8357the human brain recives 11 million bits of information per secound. Times that with 12 years as generous lower limit for when humans start doing very usefull intelligence, and we are talking about 520 terrabytes of data. Several times what llms recive. And thats not counting the data recived during evolution. Now sure, a lot of this data is redundant and mostly noise, but so is most llm pretraining data..
      A baby takes months to be able to do anything besides its preprogrammed instincts like sleeping nursing and crying. And even then it is something simple like picking up toys. Arguably also things that are instinctual.
      Arguably almost nothing great, innovative, creative or intelligent is done by humans before the mid 20s. Early in a humans 23 year is when they have recived 1 petabyte of information from their senses.
      To say humans only get a few datapoints and then infer smart and novel ideas, is just plainly false. We only do that after we have been exposed to terabytes of information about the world around us. And we only do it well after petabytes.

  • @SarahBoyd1
    @SarahBoyd1 Місяць тому

    Francois Chollet is a hero for assigning these limitations to compute in the evaluation for the ARC prize. Efficiency of inference is the thing.

  • @sampatel90069
    @sampatel90069 7 місяців тому +4

    I am beyond fully convinced that this is the best podcast on the whole of the global internet.

  • @HanqiXiao-x1u
    @HanqiXiao-x1u 5 місяців тому

    It’s a very engaging conversation, clearly the host is very passionate about the topic and excited to converse

  • @hinton4214
    @hinton4214 6 місяців тому +4

    Dwarkesh, don't listen to the comments, you did extremely well in this interview, much better than Lex when he challenges his guests. Well done and continue this line!

  • @que_93
    @que_93 6 місяців тому

    The section of the video from "Skill and Intelligence" to the end of "Future of AI Progress" made for such a compelling and engrossing watch. I must convey huge thanks to Dwarkesh for pressing on the matter-- with such vast grey-area (between "memorization" and "intelligence" (AGI))-- given the present state of research-- that allowed Francois to give out so many priceless points which can rarely be found elsewhere. Francois's insight and experience coupled with Dwarkesh's active back-and-forth questioning gave some insights which are extremely valuable, at least for someone like myself. And I have to commend his impeccable presence of mind, as at 47:58. If it not had been this active, it would have been a bland conversation (podcast). The questions assisted in giving out so many brilliant viewpoints by Francois.
    Those commenting on Dwarkesh's fast speaking style and Francois's slow paced speaking, it is their individual style. He (Francois) spoke in the same pace with Lex, and Lex himself is a slow-spoken man. It is one's individuality. With his same fast-paced speaking, he made the conversation with Sholto and Trenton very engaging and worthwhile, and so enlightening, as this one. And if I am not wrong, Francois wouldn't be this slow in uttering if he was rather speaking French.
    In the very beginning of this interview, Dwarkesh told that he had so many guests who were strong proponents of LLMs and their scalability-- which he found a bit "contrived." So, his attempt here is clearly to use those same arguments, some of his own viewpoints, and present the questions to Francois for a better and clearer insights for us, the listeners and the learners, as to what are the "bottlenecks" of LLMs which he perceives-- given there massive "accomplishments" in these last 4-5 years, and what different mechanisms or tests are their to achieve steps closer to "true AGI." This was the backdrop of his questioning. And he created a brilliant premise for such a learned and esteemed guest of the field.
    Had he not presented a case for LLM while Francois presented a case "against" LLM, it would have been a one-sided talk-- just asking questions on ARC. How else would we have come to know as to why LLMs haven't fared as well on ARC these past 5 years while they have done tremendously well on other benchmarks? How else we would have gotten so many precious inputs from Francois?
    I bow with respect to Francois for understanding the reasoning behind the premise set for this interview, and to Dwarkesh for his brilliant questions, "hard-presses" on so many important points, and for his presence of mind. I got to learn SO MUCH from this.

  • @gunnerandersen4634
    @gunnerandersen4634 7 місяців тому +16

    I think it's not hard to understand his point: Ask GPT to create a novel architecture for AGI, it can't because it can't comeup with an actual novel thing, but a mix of existing ones that looks "novel" but it really is not.

    • @falklumo
      @falklumo 7 місяців тому +5

      That's not the issue. The issue is that GPT can't think about the problem. Even humans come up with new ideas by recombining existing bits and pieces. Look at Jules Verne's "rocket" to the moon which was a big big cannon ball ...

    •  7 місяців тому +2

      Not actually true.

    • @gunnerandersen4634
      @gunnerandersen4634 7 місяців тому +1

      @@falklumo okay but it's not the same to combine things to make a new thing than actually proposing a completely novel idea right? So what I understood from his claim was that actually models currently use what they already know and combine that, but they can't take a broader look and redefine, let's say quantum mechanics with a truly novel approach, isn't like that? Maybe I got that wrong IDK 🤔

    • @gunnerandersen4634
      @gunnerandersen4634 7 місяців тому +1

      I might have missed something then, I just thought that's what this test was all about, perform over completely unseen problems that require to actually abstract more from the details and into the bigger picture to try and find the underlying logic of it. I am not an expert on any of these things, I just thought that was it.

  • @paulofalca0
    @paulofalca0 6 місяців тому

    One of the best AI talks I've seen in the last months,thumbs up for Francois Chollet and Mike Knoop. Also thanks for
    Dwarkesh Patel bringing them :)

  • @cojo9426
    @cojo9426 7 місяців тому +13

    Interviewer did not internalize anything Francois said, my god.

    • @ZachMeador
      @ZachMeador 6 місяців тому +2

      something weird going on in these comments, and this podcast episodes, where everyone is totally talking past each other. honest question: what is your explanation for dwarkesh's question on gemini understanding, and translating, the dead out of sample language?

  • @fernandoayon876
    @fernandoayon876 4 місяці тому +1

    I don't get all these comments saying that there was a lot of repeated questions. The way I see it, the subject was interesting and "tricky" enough to talk about it in depth like you guys did here, yes it would seem that you repeat the same question everytime, but the answers and explanations from Chollet were super interesting and every time we had a new way to look at it, nice interview.

  • @jeroenbauwens1986
    @jeroenbauwens1986 7 місяців тому +5

    More interviews about the importance of system 2 thinking would be awesome, for instance John Carmack (of Doom and MetaAI fame) is also working on this... Your channel is becoming so popular it could easily lead to a technical breakthrough at this point

    • @falklumo
      @falklumo 7 місяців тому +2

      The importance of System-2 thinking is now a trivial fact for everybody working on AGI. But this channel helps to popularize this.

    • @jeroenbauwens1986
      @jeroenbauwens1986 7 місяців тому +3

      @@falklumo if that's true, why did Dwarkesh' ML friends who are working on AGI not know about the ARC benchmark and why were they surprised that the frontier models failed?

    • @bazstraight8797
      @bazstraight8797 6 місяців тому

      @@jeroenbauwens1986 The German guy (interviewed in immediately previous podcast) is young (inexperienced) and not a ML researcher. He stated that he was extrapolating straight lines derived from existing data. DP also recently interviewed a pair of guys about LLMs who had only been in ML for a year ie inexperienced.

    • @jeroenbauwens1986
      @jeroenbauwens1986 6 місяців тому

      @@bazstraight8797 so your point being that people like John Schulman and Ilya Sutskever are more knowledgeable.. I wouldn't be too sure they know about ARC though, Ilya has said in the past that scaling's essentially all you need. It sounds like this might be a blind spot in all of these companies. I guess Dwarkesh missed the opportunity to ask them

  • @rufex2001
    @rufex2001 6 місяців тому

    EXCELLENT EPISODE! These types of counter arguments against the LLM hype are SUPER IMPORTANT in this public debate, and both Francois and Dwarkesh made great points for both sides of the debate! The pushback from Dwarkesh was excellent, but we need that type of pushback against the proponents of scale = AGI as well.

  • @telotawa
    @telotawa 7 місяців тому +5

    interesting thought experiment:
    if you had all the data of everything that happened in medieval economic systems, before the industrial revolution: every conversation spoken, and lots of multimodal data, trained the biggest LLM ever on all that, and then jumped forward to today: how well would it do?

    • @benmaxinm
      @benmaxinm 7 місяців тому +2

      I like this!

  • @YuraL88
    @YuraL88 7 місяців тому

    The channel definitely deserves more subscribers) Extremely interesting discussion, and I'm looking forward to new interviews.

  • @MIKAEL212345
    @MIKAEL212345 7 місяців тому +15

    this video was basically a philosophical debate on "what is intelligence?"

    • @falklumo
      @falklumo 7 місяців тому +3

      Not philosophical, scientific. A nice example where philosophical discourse stalls and scientists just move ahead :)

    • @MIKAEL212345
      @MIKAEL212345 7 місяців тому

      @@falklumo this is 100% philosophy. Science cannot answer what is intelligence. Only when philosophy gives a definition of intelligence, can science go out and measure whether and how much beings are intelligent.

    • @AlessandroAI85
      @AlessandroAI85 7 місяців тому +5

      No it's not, it's about the difference between acquire new skills and reasoning on new situations on the fly. Only the interviewer doesn't seem to grasp the difference!

    • @MIKAEL212345
      @MIKAEL212345 7 місяців тому +2

      @@AlessandroAI85 yes it is. They are both trying to define intelligence and find the boundaries of what is and isn't intelligence as compared to just memorization. What else to call that but philosophy? Philosophy is not a special thing only done in universities.

    • @YuraL88
      @YuraL88 7 місяців тому

      ​@@AlessandroAI85 they both had hard time to understand each other. The main issue is emergent properties. You can't predict if a large enough model have some "true" intelligence or llms are saturaing.

  • @YuraL88
    @YuraL88 7 місяців тому +1

    It's such a fascinating discussion! It definitely deserves more views! I can disagree with Francois Chollet about LLM potential in general, but I must admit that his approach is extremely refreshing and novel. We need more people with nontrivial ideas to build True AGI because just scaling LLM is a risky approach: if we fail the new great AI winter is waiting for us.

  • @lra
    @lra 6 місяців тому +6

    This dude actually threw out the line "I studied Computer Science" to Francois Chollet. We get it, you're an LLM fanboy, but you're speaking to someone that actually knows what they're talking about. Let the man speak!

  • @tusharbhatnagar8143
    @tusharbhatnagar8143 7 місяців тому +2

    I have 2 questions for everyone present in the video...
    1. If LLMs are able to perform economically valuable tasks through memorization and interpolation, does it matter if they aren't "truly" intelligent?
    2. Can the advanced tooling and features offered within ChatGPT, combined with self-learning LLMs, be viewed as achieving the form of intelligence referenced in the video? (Can be any AI offering, using ChatGPT merely as an example)

  • @cafer12098
    @cafer12098 7 місяців тому +15

    I dont understand why hyperscale people are so stubborn…. It is obvious this architecture does not possess the ability to reason… Lets scale it up and use it but lets also spend equal resources on finding new architectures…

    • @francisco444
      @francisco444 7 місяців тому +3

      But it does reason... otherwise we would be stuck with GPT 2 level intelligence

    • @BadWithNames123
      @BadWithNames123 7 місяців тому

      Lol. Have you ever tried paid models?

    • @JumpDiffusion
      @JumpDiffusion 7 місяців тому +1

      “I don’t understand…”. Yeah, we see that…

  • @d96002
    @d96002 6 місяців тому +1

    Very interesting conversation !. I'm not an espert but it definitely feels to me like as soon as you step out of the path paved by memorization the current Ais give incorrect answers, look for example at the responses of Gemini about pizza and glue etc. Unless they want to make us believe that it is only memorization...

  • @1purpleman1
    @1purpleman1 7 місяців тому +8

    Love the podcast but Dwarkesh essentially just kept trying to push the same argument repackaged 150 times.

  • @dragonystic
    @dragonystic 7 місяців тому +2

    I don't find Francois's opinions very compelling.
    I don't think humans have general intelligence the way he thinks we do. At 24:15 he mentions general intelligence means being able to learn new skills very quickly. I mean, can humans learn new skills very quickly? We have to spends years in school to get many skills. If we wanted to learn a musical instrument, again, many years of training. And that's the best case scenario!
    The human brain has the capacity to learn. And so do these LLM systems. I don't see much of a difference.
    He also makes a big deal about novel situations. Saying humans can approach a novel situation and adapt. I don't think this is correct either. A human is multi-modal, so we have a lot of data we are trained on. There really aren't that many truly novel situations. LLMs are trained only on text. So it's not surprising that the Arc test can come up with questions that require a kind of data it hasn't come across, and it thusly does bad.
    Once these systems are truly multi-modal, I have a feeling they will do quite well on these kinds of tests.

    • @wwkk4964
      @wwkk4964 6 місяців тому +1

      Yes, if we hand the ARC test to piraha tribe of the Amazon or Jarawa Tribe in Andaman islands, they will not pass it. I can guarantee Francois will not say they are just memorizing patterns and not intelligent.

  • @vincentcremer4235
    @vincentcremer4235 7 місяців тому +5

    Dwarkesh is totally right pressing on the critical claim that humans do more than template matching. Chollet couldn't give a single example. The question was very easy to understand.

    • @power9k470
      @power9k470 7 місяців тому +6

      You think the discovery of quantum mechanics and general relativity was template matching. You know nothing about the capability of the human mind.

    • @vincentcremer4235
      @vincentcremer4235 7 місяців тому +1

      @@power9k470 there is a structure to this world that we seem to be able to partially observe. We create models of this structure. Whatever model we create - the model structure can be understood as a template. And many theories share similarities. One could think of the application of modus ponens/ deduction as one such template.

    • @vincentcremer4235
      @vincentcremer4235 7 місяців тому +2

      @@power9k470 and it's also not about who is right. I am not sure. The problem is just that Chollet claimed there are many examples but couldn't name a single (besides his test).

    • @power9k470
      @power9k470 7 місяців тому

      @@vincentcremer4235 Chollet probably could not because he is a CS and ML guy. These subjects do not have profound paradigm shifts. This question is suitable for physicists and mathematicians.

    • @benprytherch9202
      @benprytherch9202 6 місяців тому +2

      Think about art. Generative AI cannot create art in new, never seen styles. If you trained a generative art AI on all the paintings up until 1850, do you think it could develop new styles as original as impressionism or surrealism or comic art or graffiti? If you trained a generative music AI only on music recorded before 1960, could it come up with new styles as original as punk or hip-hop or EDM?
      Art is constantly evolving. Sure, individual artists do much that is derivative. But the fact that art evolves demonstrates a human capability that cannot possibly be a form of template-matching.

  • @johntanchongmin
    @johntanchongmin 7 місяців тому +1

    "You cannot navigate your life using memorisation" - Chollet 29:05
    My take: We need some form of memory matching for decision making. We definitely do store something in memory, but not the exact memory of the event, but in different abstraction spaces.
    The benefit of different abstraction spaces is that you can mix and match various modular blocks together, and easily construct something new with this. So, even if you have not encountered a situation before, you can have some idea of what to do based on something similar that you have encountered before.
    And the thing about human memory is that it need not even be an exact replica of what you have experienced - it can change based on retrieval, but as long as it is still relevant to the context, at least you are starting from something.
    So, memory is definitely the way we do decision making.

  • @Limitless1717
    @Limitless1717 7 місяців тому +9

    Francois is arguing what, that Humans and LLMs think differently. Yep, we know that. But what he doesn’t want to acknowledge is that LLMs (ie ChatGPT 4O) have some degree of reason and inference that seems to grow organically through brute force modeling of language alone.

    • @RogerJL
      @RogerJL 7 місяців тому +3

      But ChatGPT 4o reasoning is weak. When you detect a fault in an answer, ask it to check that fact - it does and comes to the conclusion it was wrong
      Even lists the reason for why it was wrong!
      Now ask it to reprint - the correction is gone... (still being in the same session)

    • @Limitless1717
      @Limitless1717 6 місяців тому +1

      @@RogerJL The fact that it can reason - at all - suggests that brute force learning, if large enough, could still arrive at AGI.

    • @RogerJL
      @RogerJL 6 місяців тому

      @Limitless1717 but does it really reason, or is it just replying with the most reasonable response to a human input about it being wrong?

  • @shyama5612
    @shyama5612 7 місяців тому +2

    This is really insightful from francois - great articulation of memorization/benchmarks gaming, true generalization, intelligence and a good working definition of intelligence from Piaget - I recall the other frenchman yann quote him once too. I'd love to see the day when google/Francois ' team creates a new System 2 based Combinatorial search (discrete search) based engine that can do program synthesis on the fly with LLM embedded in them in the future!

  • @sloth_in_socks
    @sloth_in_socks 7 місяців тому +3

    Prediction: Arc will be solved within a year. Dwarkesh will claim he was right because it was a scaled up model. Francois will claim he was right because it wasn't purely an LLM. In the end it doesn't matter because we got closer to AGI.

    • @eprd313
      @eprd313 6 місяців тому

      This, and people will still be stuck in their reactionary prejudices or desperation to see change