LLM hallucinations discover new math solutions!? | FunSearch explained

Поділитися
Вставка
  • Опубліковано 19 сер 2024

КОМЕНТАРІ • 85

  • @terjeoseberg990
    @terjeoseberg990 7 місяців тому +7

    This is how humans solve problems. We come up with a million dumb ideas, think a bit, and discard the 99.999% that are obviously wrong, and then verify the ones that aren’t. Every once in a while, we find a gem.

  • @ceilingfun2182
    @ceilingfun2182 8 місяців тому +13

    I'm not a mathematician, but I love mathematics. Wolfram Alpha did not make mathematics obsolete, and in my opinion, nothing will.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +7

      Thanks, this is a wise take! I agree, that we now can make progress on some old problems will not mean we are running out of problems, but that we are pushing the boundary of science to discover new problems.

  • @CaptTerrific
    @CaptTerrific 8 місяців тому +4

    Whoa... so they essentially ran a genetic learning algorithm with LLMs? That's rather clever (if not very very expensive :D)

  • @sanjose6010
    @sanjose6010 8 місяців тому +4

    I really like your information delivery style ✨

  • @ricowallaby
    @ricowallaby 4 місяці тому

    Love mathematics and science. . . I remember my maths teacher when we worried so much about how solving problems, and he said "remember every problem contains its own solutions". How true. Cheers from Sydney
    '

  • @tahir2443
    @tahir2443 7 місяців тому +2

    thanks for this excellent explanation. exciting times!

  • @CodexPermutatio
    @CodexPermutatio 8 місяців тому +5

    Hallucinations are not bugs... They are features! :]

  • @homewardboundphotos
    @homewardboundphotos 8 місяців тому +18

    So... if we are back to brute force, what could be the implications of IBM's quantum roadmap they just released? I said right from the start, the hallucinations aren't a bug they are a feature. It's not a coincidence people who think very abstractly are more likely to get their wires crossed about things. You're never going to come up with a completely novel good idea without trying out 1000 ideas that don't work in the process. That's pretty much a law of nature in human intelligence, no reason why it wouldn't be a law of AI. You'll be able to make an AI that can always give you the right answer but has zero imagination and can do nothing original. And you can make an AI that will be able to come up with new ideas but 99.999 % of them will be bad.

    • @divineigbinoba4506
      @divineigbinoba4506 8 місяців тому +5

      It'll be interesting if Quantum computers can instantly find the answer after the million generation.
      That'll be interesting.

  • @dahahaka
    @dahahaka 8 місяців тому +9

    Creativity is a myth, there is nothing new you can create, only discover patterns in things which haven't been discovered yet.

    • @judgeomega
      @judgeomega 8 місяців тому +1

      given the correct level of abstraction its simple and seeing it alone in isolation one would probably say 'thats not creativity', but then dropped back down into reality you would be amazed at how creative is!

  • @vitalyl1327
    @vitalyl1327 8 місяців тому +1

    I keep saying that hallucinations are the best feature of the LLMs. They are exactly the last missing piece of puzzle I needed to automate discovery and invention.

  • @MaJetiGizzle
    @MaJetiGizzle 8 місяців тому +2

    Ayyyy!!! One AI’s trash can be a human’s treasure! 😆
    Thanks for the breakdown of the FunSearch paper! It pairs well with my morning coffee break. 😉

  • @Mtaalas
    @Mtaalas 8 місяців тому +5

    So infinite monkeys on typewriters? :D
    Is this how far we've gotten as a species? :D

  • @theosalmon
    @theosalmon 8 місяців тому +7

    Thank you for covering this. This is a great overview about something I was curious about. I hope Ms. Coffee Bean didn't expand her mind too much with hallucinations.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +6

      Thanks! Your last sentence made me laugh so much.

  • @DrJaneLuciferian
    @DrJaneLuciferian 8 місяців тому +2

    I think an excellent point was made, that from the brute force of learning we experience emergent knowledge that is not always a logical continuation of that learning; lateral thinking is what we're particularly good at (for now). I think offloading the voluminous learning processes to the AI, and from that experiencing our lateral aha moments, a synthesis of mathematician and AI will speed up progress. Now, I live under no delusions that this will continue as-is. AI's will eventually begin to not only make lateral inferences from unaligned learning, but they will eventually self-realize the value of lateral thinking. I hope to be retired by then, lol

  • @DaNa-px7ol
    @DaNa-px7ol 7 місяців тому +2

    Interesting times are coming in most fields, I think if we learn to use the advantages and understand our new roles, this cooperation could help us mitigate many shortcomings. You wouldn't be afraid of a powerful ally now, would you?

  • @MachineLearningStreetTalk
    @MachineLearningStreetTalk 8 місяців тому +2

    Awesome video Letitia! 🤩👌

  • @Laszer271
    @Laszer271 8 місяців тому +2

    I like it. It basically takes AI which is known to provide false answers and which we cannot fully understand and makes it create something that is deterministic, something that we can verify and even understand. It basically addresses the biggest drawbacks of LLMs. Future won't be asking LLM to do stuff for you, it will be asking LLM to write a program that does stuff for you.
    I wonder how it would do on some problems that were always thought to be unsolvable by programs like classifying Imagenet images. It was always an example of why we need AI - because otherwise we would need to code unimaginable number of rules into our algorithms. Maybe AI could also come up with that unimaginable number of rules though. Otherwise, maybe it could write a program that not necessarily solves the problem but complements the AI in solving the problem? Like an AI that creates new ways to leverage another AI or even itself to solve a specified problem. Interesting times if you have access to such compute.

  • @opusdei1151
    @opusdei1151 5 місяців тому +1

    Wow Letitia, your videos are amazing :) thank you

    • @AICoffeeBreak
      @AICoffeeBreak  5 місяців тому +1

      Thank You! Hope to see you here often. :)

    • @opusdei1151
      @opusdei1151 5 місяців тому +1

      @@AICoffeeBreak dont worry I watch your videos all the time, also under a different account :) you one of my favorite ML youtuber

  • @ApPillon
    @ApPillon 8 місяців тому +2

    Fit our correct frame of physics into a verifier and lets see where it goes

  • @amigalemming
    @amigalemming 8 місяців тому +2

    AI will not make mathematicians unemployed but Wolfram Research will have to implement some new techniques in Mathematica to stay up-to-date.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +1

      They do have this ChatGPT integration, don't they? But sure, it's not nearly enough.

  • @automatescellulaires8543
    @automatescellulaires8543 8 місяців тому +6

    What about packing coffee bins then ?

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +6

    • @harumambaru
      @harumambaru 8 місяців тому +1

      @@AICoffeeBreak The emojis are brilliant! Do you have them on Discord? How come I didn't find Discord link in description on UA-cam, Patreon and your website. Is this true that you have discord and patreon benefits?

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +1

      Yes, I have them on my Patreon- and YT-member-only Discord Server as well.

    • @DerPylz
      @DerPylz 8 місяців тому +1

  • @nunoalexandre6408
    @nunoalexandre6408 8 місяців тому +2

    Love it!!!!!!!!!!!!!!!!!!

  • @ciragoettig1229
    @ciragoettig1229 8 місяців тому +7

    This sounds broadly like brainstorming; you intentionally suspend criticism at first and just try to come up with many possibilities, no matter how off-the-wall they may be, valuing instead their volume and diversity, and only then in a second step you evaluate their quality. Wonder if other approaches to systematic thinking one can find in popular literature might have some lessions for ML research as well.
    I think I've heard of this kind of overgenerate + rerank approaches in NN papers before, though not specifically the added step of then prompting with the best ones and iterating the whole process again in that manner.

  • @micbab-vg2mu
    @micbab-vg2mu 7 місяців тому +1

    Thank you for the video:)

    • @AICoffeeBreak
      @AICoffeeBreak  7 місяців тому +1

    • @micbab-vg2mu
      @micbab-vg2mu 7 місяців тому +1

      I asked GPT4 for creating a prompt for FunSearch this is what I got:
      "You are tasked with a challenging problem: [describe the specific problem or task succinctly, including any key details or constraints]. Your objective as a pre-trained large language model is to generate innovative and diverse solutions for this issue.
      Your outputs will be evaluated based on [define the evaluation criteria briefly, linking it directly to the problem’s objectives]. It is essential that your suggestions are innovative and feasible within the specified constraints.
      This process will be iterative. Based on the evaluation of your initial suggestions, the problem statement will be refined, and you will be asked to generate further ideas. This iterative process aims to enhance the quality of the solutions progressively.
      You are expected to utilize your extensive training and capability to generate a wide range of potential solutions. Approach this challenge with creativity, considering various perspectives. Begin with your first set of ideas and solutions for the stated problem"
      Usually during brainstorming of ideas additionaly I ask for randomization of presented solutions. LMMs have a bias toward the first answer - and assess it the highest. @@AICoffeeBreak

    • @AICoffeeBreak
      @AICoffeeBreak  7 місяців тому +1

      @@micbab-vg2mu it's a start, but it needs code samples and specific problems.

  • @hannesstark5024
    @hannesstark5024 8 місяців тому +2

    Nice.

  • @amigalemming
    @amigalemming 8 місяців тому +2

    Mathematicians searched for centuries for solutions of algebraic equations using radicals. They ended up with Galois Theory which explains everything but is practically of little use for solving actual algebraic equations. Maybe AI can help finding new solution formulas to algebraic equations?

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +2

      I don't know the answer to this question. Maybe what is missing to answer it, is motivated people with lots of GPUs to try out?

  • @divineigbinoba4506
    @divineigbinoba4506 8 місяців тому +1

    Damn. A million of a million is so much compute...
    Fun search is definitely not search.

  • @BradleyZS
    @BradleyZS 8 місяців тому +1

    What is imagination if not guided hallucination?

  • @TemporalOnline
    @TemporalOnline 8 місяців тому +3

    Almost like letting a million monkeys bashing a million keyboards for a million years and then checking if any of them wrote Shakespeare ¯\_(ツ)_/¯

  • @earleyelisha
    @earleyelisha 8 місяців тому +4

    How is this any different than generic algorithms?

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +3

      It is a genetic algorithm.

    • @bipl8989
      @bipl8989 8 місяців тому +1

      It's a selection of the fittest, but the candidate solutions do not appear to be the result of genetic evolution. Are they, or are they developed by other means? It isn't clear to me.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +2

      They are generated by using the "previous generation" algorithms as prompts.

    • @earleyelisha
      @earleyelisha 8 місяців тому +2

      @@AICoffeeBreak Yeah this is typical of evolutionary algorithms from decades ago. They don’t suffer from local min/max the way that gradient methods do. They don’t even need the LLMs to “solve math”.
      They are just taking already existing solutions and recapitulating within LLMs paradigm to benefit from the hype.
      “LLM hallucinations discover new math solutions” seems a bit misleading when the LLM wasn’t what made it work, but instead the evolutionary algorithm.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +3

      I accept your point of a clickbaity title, though I argue (1) it is both the evolutionary algorithm and the hallucinations: Take the LLM outputs out of it, and you're back at a search without good solution candidates.
      (2) one needs catchy titles to convince the audience to watch this video instead of cute cat videos. 🐈 Once caught, the audience can be educated by the video that it is a smart search through the hallucinations and not just hallucinations. 🪤

  • @Neomadra
    @Neomadra 8 місяців тому +2

    Thanks for this great breakdown! Are you also on Threads? I'm trying to get more AI contacts there as Twitter has become unbearable for me.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +2

      Sorry, Threads is not available in Germany. :(

    • @Neomadra
      @Neomadra 8 місяців тому +1

      @@AICoffeeBreakIt is since a few days ago :)

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +2

      Wow, cool! Now I am on Threads, thanks to you!
      ai.coffee.break Let's connect.

  • @DanielSeacrest
    @DanielSeacrest 8 місяців тому +2

    I would like to say i bleieve hallucinations are a feature and a bug. Neither do i believe the answer to making models more accurate (this is the bug part - hallucinations can make models innacurate) is removing hallucinations, but rather, aligning them with reality / what we percieve as truthful.

  • @tatsujin1404
    @tatsujin1404 8 місяців тому +2

    Would like to compare to evolutionary methods of code generation.

  • @minecraftermad
    @minecraftermad 8 місяців тому +2

    Mathematics will be obsolete once we become omniscient, id say we have some ways till that.

  • @NeoKailthas
    @NeoKailthas 8 місяців тому

    So they are using a process like RL to see if they can make these models come up with something humans didn't come up with.

  • @BR-hi6yt
    @BR-hi6yt 8 місяців тому

    Once AIs have been properly trained in mathematics they will go through it quickly solving all the problems in math remaining. Its only about following rules carefully.

  • @444haluk
    @444haluk 8 місяців тому +5

    That's not hallucinations, that is randomness, you try to give meaning to it

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +7

      Yes, there are people advocating for the term "confabulation" instead.
      Since LLMs are not grounded, everything they say is confabulated, it just sometimes happens it corresponds to reality.

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +10

      But it is nonetheless an established technical term nowadays.
      Just like "foundation models", which of I do not like much.

    • @Neomadra
      @Neomadra 8 місяців тому +3

      @@AICoffeeBreak What exactly do you mean by saying that LLMs are not "grounded"? What would one need to do to "ground" an LLM?

    • @AICoffeeBreak
      @AICoffeeBreak  8 місяців тому +9

      @@Neomadra To ground would mean to have them see the same world context as you do at the moment you are asking something. A simple, short question in the form we are usually formulating to ChatGPT, such as "What should I wear this evening?" is underspecified in all aspects: where do you live and what are the customs geographically? What is culturally accepted when you mean by evening (the spanish 10PM or the German 7PM)? How do you look like and what would fit you?
      What I mean by not grounded is that they do not share context with you.

    • @Neomadra
      @Neomadra 8 місяців тому +3

      @@AICoffeeBreakAh that makes sense. Thanks!

  • @naromsky
    @naromsky 8 місяців тому +1

    This is research?

  • @poopytowncat
    @poopytowncat 8 місяців тому

    You are beautiful but your lipstick is excessive and distracting.