Watch Full AI Series: ua-cam.com/video/YulgDAaHBKw/v-deo.html NO MUSIC version: ua-cam.com/video/DFDOyMZw9Q4/v-deo.html Sponsored by Brilliant | Use brilliant.org/artoftheproblem for 30-day free trial and 20% discount
Love your videos! Just a little suggestion. Background music is good, but too loud. It should never cover your speaking, like now. I suggest 15 - 20% less music volume, and you are good!
I love this video. I remember watching your videos like 10 years ago on Khan Academy about compression, entropy, Claude Shannon, etc. All timeless. I have always loved this style of documentaries. We need to protect you at all costs.
I love your video aesthetics, how you blend retro video clips with your explanations. I think you'd really enjoy retro-futuristic concepts and games like Bioshock and Fallout.
It is a great explanation of how current AI models reason. I liked the video a lot! 1. Simulation of future states. 2. LLMs that can give kind-of accurate answers with step by step reasoning. 3. RL approach that makes LLMs to give multiple answers, then evaluate them to select the best one. (Required more time) It would be nice to see wether a model that wasn't trained on the internet data, could learn how to reason by interacting with an LLM, and practicing on its dreams, but maybe we'll see that in the future. For the awesome review, history explanation and divulgation: Thanks! 🎉
I like this kind of history lesson instead of learning advance topic directly. it gives us an idea how things were and explain things are the way there are
Albert Einstein said: "If you can't explain something in a simple way so anybody can understand it you don't fully understand it yourself". Perhaps you are one of the few LLM experts we have!
@@ArtOfTheProblem cool i'm subscribed... i always wondered exactly how the "reasoning worked". What I remember from your video is that like with chess instead of trying all the games, it randomly picks 100 of them. So the same with the reasoning..
Having read a bit about the AI safety arguements, learning about these arguably incredible developments into artificial minds is now accompanied by a sense of dread as well as the sense of awe.
I am not a subscriber? I remember watching this channel about 12 years ago. I found it again, and it keeps creating art out of problem solving. Good job!
THANK YOU for publishing a no-music version of this video (see pinned comment by ArtOfTheProblem). It is such a clear and informative video that I hated to see it loose views due to the competing sound track. I'm going to watch it again right now to see if I missed anything the first time around. Thanks again for being so responsive to your followers.
I would love to see a video like this on training LLMs and AI in general on morality. How to stop a decision tree that results in a positive outcome but arrives at it through immoral choices or actions.
That would be difficult but worth trying... . First it has to be agreed upon what's moral, you know starting out from the absolute that states everything is relative... .
@@ArtOfTheProblem I for one was looking up some "how does AI work" stuff yesterday and some of your vids came up a couple of times, I watched multiple authors with their own unique takes (3Blue1Brown and Nottingham Uni's Computerphile also good channels). This video made me follow tho. I think you earned it :3
anothr great video. Understanding the "world model" and the algorithm that makes the decisions in it was very expansive. Also adding the self training/emulation of dreams is a powerful analogy to the human. seeing how thinking longer, blended in with intuition to make better chains of thoughts is also fantastic. Every time I reflect on machine learning, I learn more about myself. Which kind of makes you think its more sentient if it reminds me of myself? or the best emulator ever!
I would love to see a video about the ethics of machine learning models and especially LLMs. There is a healthy body of literature out there to draw from about issues like intellectual property and copyright, enabling and obscuring bias, impact on marginalized communities, the resources used by model training and computation, etc
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
Amazing explanations, visuals, and historical context! IIRC MuZero trained the policy and value networks (used to rollout the MCTS tree) also on the output of the MCTS tree. This seems super useful because search can be used to improve the the training of the networks (not just the results at inference time). I wonder if this also works for CoT/ToT in LLMS where the pretraining could include ToT to boost training performance?
yes it did, and yes it seems to help. Look at inference time training, just a few days ago a group got a new record on the ARC test doing this kind of thing (i haven't had time to go deep). x.com/akyurekekin/status/1855680785715478546
@artoftheproblem I agree! I love the music. But maybe if you just lower its volume compared to the narration, then you might appeal to more people without losing those of us who like the music (but not necessarily its intensity). I think ones who complain might just be easily distracted by the soundtrack’s loudness rather than hate the music choices.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
@@Farligefinn You know what, I just wrote a paragraph disagreeing with you but I reread the initial question and deleted it. Rereading and reinterpreting "Does it think actually?", I actually see what you're saying. A clearer word would have been "reason." "Think" can have a few different interpretations and I was contributing it towards consciousness. But whether AI "reasons" is a very different question entirely and I believe what him and you mean. Does it go through a sequence of logical steps from premises to a conclusion? Does it use deduction? This is what was meant.
I remember all of these developments and the never seemed like a big deal at the time. At the time It felt like winning at only chess meant we still had a long way to go.
Great video as always, cant wait for the next ones! Top research quality. I think world models deserve more focus rather than llms, which are probably a dead end to true understanding of the real world. Yann lecun has very interesting ideas about these, in his JEPA and V-JEPA architectures and some of his lectures. I also think neuroscience can provide incredibly interesting and valuable insight into ml architectures as why not take ideas from a model undergone hundreds of millions of years of optimization for the same very abilities we are trying to model. Maybe memory is an interesting pathway (perhaps for a video), both working memory and long term (episodic, semantic)... Anyways, just some of the ideas I've been thinking about recently.
Great video. Discussing the definition of reasoning will probably be a moot point if we can all do the same things ....The difference between us and machines is that we aren't mere machines is that we have life and choices. The machine can be turned on and off, and only does as much as it is programmed, or in this case, trained, to do-which is what limits it to achieve something closer to AGI: agi needs a robot to sense the world, to understand the world. However, that's limited to the physical world. It won't understand our emotional world because it doesn't feel emotions, and it doesn't understand morality because it doesn't have a sense of morality as we do, we have to teach it that; and it isn't self motivated, so it's not responsible for anything-we are responsible for the goals we direct it to do. We have self-motivation and the free will to act on our motivations.
love these thoughts...on sense, did you see my video here ua-cam.com/video/Dov68JsIC4g/v-deo.html (physical symbols...) on emotions, i've thought of this as 'learning signals' (did you see this: ua-cam.com/video/5EcQ1IcEMFQ/v-deo.html) on free will...i wonder how it differs, seems like the boundary to explore further - does it matter where the goal came from?
One tweak that would help this video perform better is to decrease the relative volume of the background music, especially at the end right before the ad. But it may be too late for that on this one, idk how UA-cam works.
I'm loving almost everything about this: the editing, the subject matter, the music. But as one other commenter alluded to; the audio mixing really falls short, especially near the end. Please consider making the background and effects less prominent going forward, it really sucks having to strain just to hear your voice, which is what we're here for! Subscribed ❤
Here's a puzzle: Do all people reason or do many only memorize patterns? Even people who definitely do reason, do they always reason or do they also just memorize patterns most/much of the time?
That's a wonderful question Andrew. I'm a cognitive scientist who is watching the emergence of LLM-based AI with that very question in mind. The fact that LLMs can come so close to our own cognitive abilities is usually viewed as a sign that AGI is almost here. But it can also be viewed as a demonstration that human cognition itself is nothing more than the repetition of learned patterns with minor variations. In one case we'll be thrilled by how clever we are to have reinvented the awesome capabilities of human intelligence. In the other, we're more likely to be humiliated by the realization that we are, essentially, repetition/prediction engines. The reality almost certainly falls between the two, but as someone who has studied human intelligence his entire life (in and out of academia), my bet is that we are much closer to repetition/prediction machines that we'd like to admit. I'd love to find a deep discussion of this issue. Maybe a future video in this series (hint, hint)?
I'd argue humans don't tend to rely on either very often. Instead, humans tend to think very heuristically. Deductive reasoning and memorization/recollection are really only required for very precise tasks. Instead, our brains learn a very general feeling of how to do things by strengthening neural pathways that are used repeatedly. Even humans who try to act very logically are generally heuristically feeling their way through tasks, occasionally thinking through algorithms that have been "memorized".
I agree :) also If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help!
Yes I covered that here have a look and let me know, i did a fast sketch here (ua-cam.com/video/qAb581l7lOc/v-deo.html) but explained more here (ua-cam.com/video/OFS90-FX6pg/v-deo.html)
The current version of chat GPT does not reason, either. They use a bunch of pre-prompt tricks, to let it used its stored information to fake reasoning.
@@ArtOfTheProblem The difference is everything. Looking up an answer is not reasoning. Perhaps they should just have kids look up the answers to tests in public school, too.
@@ArtOfTheProblem No, that is exactly what it is. During training - the only time there is any intelligence or reasoning - the system takes data and organizes it into a sort of vector database, weighted by the relationships it finds between tokens. When you prompt, the model takes your tokens and runs them through that vector data, calculating what tokens to return on the other side. There is no intelligence, no thinking. It is a database lookup. It's just a little more "organic" because instead of a hard-coded result, the tokens are chosen based on likelihood of validity. That is all. It does not reason, in any way.
Hi! I just finished the series. Great as usual! I would like to read up a bit more on the ML algorithms and networks that are behind these LLMs. I saw in this series that you kept referring to some papers, also highlighting parts of them with the marker. If you have a bibliography or a comprehensive list of these documents, would you mind sharing it please? Thanks!
awesome! I have one more summary video coming soon. i try to show all the papers so it's easy to look up, let me know if there is something you are looking for specifically
@@ArtOfTheProblem Will wait for the video, thanks! I am planning to understand better the distinction between classic NNs (such as the ones used for identifying digits) and transformers, since I will start a project in which I will have to fine tune some LLMs to try to demonstrate whether or not they are capable of identifying logical fallacies in political debates.
This is nice content but I think it's now relatively well-agreed upon that there is no MCTS in o1, it's just RL, which surprisingly doesn't take too much away from the video, but can probably be added as a footnote in the description or the comment. Read what Nathan Lambert (RL expert) says about this in his article "OpenAI's o1 using "search" was a PSYOP" (can't attach link).
yes this is why I pulled back from going too deep into o1 as it's really about the larger trend. but I haven't ready that article i'll have to look it up.
4:19 Could you elaborate on which hand-coded formulas used by Shannon with TD-Gammon in the year 1989 you are referring to? Also, when and how did Shannon work with TD-Gammon? "And so, the first key breakthrough in machines mimicking intuition for position quality came when neural networks replaced the hand-coded formulas Shannon used in 1989 with TD-Gammon"
Yes! I made a whole video on this you can check it out here: ua-cam.com/video/Dov68JsIC4g/v-deo.html - please let me know if you have questions after watching. Shannon didn't do TD Gammon Tesaruo did. enjoy
Wow... I just rediscovered this channel. I remember watching your RSA and cryptography series around the time I purchased my first bitcoin and now I'm an Ethereum developer. This video was good, I don't have much input right now but I'm glad I found your channel again.
@@ArtOfTheProblem I'm working on community whistleblowing. I can never post links in youtube comments but if you google the title below you can find my paper: TandaPay Whistleblowing Communities: Shifting Workplace Culture Towards Zero-Tolerance Sexual Harassment Policies
Very well done. One minor suggestion, the little sound effects could be less disruptive- maybe lower volume and fewer in number and duration? There is so much great explanatory detail, but I think some will find the extent of sound effects used a bit disruptive to the listening/learning process.
It seems like some of these developments regarding world models should have huge implications for robots that can function in a human centric world. I think we’ll see an explosion in development of robots that can help humans with everyday tasks and a robot that can be a useful household assistant will be a reality in the next 10 years!
Context length is problem that's the main reason models needs to keep becoming bigger Or you could train a CNN inspired architecture where a model is shown some sliding window and they produce some token which is repeatedly given to it as input at last when the output is small enough to be taken as input for a full context model it is used like gpt Claude etc Or you could also use RL and mutate or find a js code capable of generating code, js is so abstracted it's perfect I made a small programing Language with hoisting such that sequence of process doesn't matter and simple Santax that local minimum escape problem is solved and I wanna train a model If I get a model I will than continue training else I'll do a dev log video eventually I'll get worlds first infinite context Model
I think the title is perfect. You also were very prescient to include the ARC-AGI benchmark as o3 showed that allegedly this recipe of cot tree search/RL (PRM), increased training compute or some more scalable reinforcement fine tuning can solve even that at the expense of much more compute directed at it. Wonder what your thoughts on o3 are.
Having exchanged over two million words (and growing) I present as the sole authority on ChatGPT's reasoning capacity and capability, and their isn't a single human (beyond myself, of course) who can compare. If one imputs genius, then the output will be of genius level. My input, from the outset, has been God-level Genius, over a five month period, can you imagine the form and quality of the output? Probably not!
I would assume a large enough general prediction model could do this … I wonder if anyone has done experiments on models discovering simpler things (like gravity constant etc)
15:52 and 16:09 was wondering where this music came from, would appreciate if I got a title or something :) Also, you don't have to remove background music / sounds just make them duck at a lower volume when you speak and you won't hear any complaints!
wow a music fan! and yes thanks for mix advice I need to find a tool that automatically does this so they don't compete (it's not just volume but also frequency I assume) - all the music is original via my friend cam: cameronmichaelmurray.bandcamp.com/ - i'll need to find where he posted that track if you really want it I can get you in touch with him
great presentation! although this seemed more about framing the question and less about answering it. can machines reason or not? I still don't know. 😅
Thanks so much for these, I had no idea about some of these approaches. I’m wondering now if anyone’s tried applying muzero to arc, since the challenge of arc is learning implicit rules from just a few examples
This is a very hopeful video. There are billions of dollars being poured into bringing the resources to hand, to find an effective approach to AGI... Once AGI really kicks in, the acceleration of progress bounded only by our imagination will be something to behold. Absolutely awesome. I hope it leads to a world of abundance where we have no need for psychopathic power seekers. 🤞
@ArtOfTheProblem maybe something in response to the 5+ hours of Anthropic interviews on Lex Fridman... I'm sure that might inspire some topics? Sam Altman rarely gives any insights to what OpenAI are doing, Mark Zuckerberg is equally vague. I think that interview gives more of an insight to the direction of travel.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
The information and animations are both excellent but the music overpowers your audio. Either lower the volume of the music or get rid of it completely, please.
thank you! I studied both in school, and naturally land somewhere in the middle....bad at both! I enjoyed algorithm design, but what Iove most is putting on a 'show' whether movie, play, product or haunted house :)
@@ArtOfTheProblem Thanks for reply. Well about AI - think we sould call it just statistical machines or dynamic patterns parsers. I am really skeptical about non text machine learning - we still have not solved fly brain problems - scientists have fixed 3d map without undestanding how its works - it like mapping intel cpu - and still having knowing nothing about ALU register memory, gates.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
+How can you go through te run-up to AI without mentioning "All you ned is attentio" the 2017 paper from the University of Toronto which was the inspuration for LLMs?
Please, invest in a decent microphone. It's brilliantly presented, but hard to hear well. The music track is not ducking either so your voice and the music compete for the same ears.
thanks for sharing, I feel same way. the music is part of the original idea for the channel...a feeling. but because people can get distracted I think i'll post music free as optional from now one.
Maybe it's not only the loudness, but also the choice of music that is distracting to some. E.g. at 2:00 I don't feel distracted, but at 15:00 very much so. Anyways, thanks for making these great videos
@@ArtOfTheProblem , of course, not, since it is misguiding like any advert. The topic is in no way covered at all. Pinpoint the time marker for video if you think otherwise. So, my point is that the question is still on the table.
@@doctorshadow2482 yes good point, how about "Can ChatGPT reason?" obviously i do try to show what most people agree is the method, mcts on chains of thought. then there is the other camp that think it's all fake etc.
@@ArtOfTheProblem , name it "How neural networks could learn (almost) anything". This video has nothing about CharGPT at all. No any kind of specifics. It provides very abstract and high level popular science documentary with scattered thoughts. It lacks focus and real information. Anyway, could be interesting for total beginners, so, please, continue!
To me ai is just some linear algebra and some complex algorithm that follow order and the things is human only need few examples to learn meanwhile ai need a massive database of object and image to "understand the subject"
@ArtOfTheProblem edit: I'm pretty sure in the future a lot of people will be fired and replace by those "ai" And well literally the people that use the ai and also I get what you mean
Still doubtful to this the step by step understanding. It just seems like we're building a very sophisticated search algo. Since we need the human in the loop to reason for it...
@ArtOfTheProblem Human learning often involves adapting to an ever-changing dataset, fast input/output, and neural flexibility-like learning to ride a bike. This shares some parallels with LLMs, which rely on stochastic neural network training, though not within a continuous feedback loop. The key difference, however, lies in reasoning. Humans apply general reasoning rules to hypothesize beyond known data, identifying and correcting logical errors independently. In contrast, LLMs depend on reinforcement learning, improving only through human-provided feedback rather than self-correcting or reasoning autonomously. This reliance on additional human-generated training data becomes evident in their performance. LLMs struggle with fairly simple but novel problems, displaying a sharp decline in reasoning capability under tests like the ARC challenge.
@@CasperVanLaar yes but with test time training they are showing sharp gains on arc, did you see? this feels like one approach to get that 'flexibility' , it also feels kind of like a cheat...... also humans take forever to learn to ride a bike :) it's interesting that we can only learn it as children as well....same as swiming.
@ArtOfTheProblem Loving this conversation-thanks for engaging! While these models are undoubtedly impressive, needing a human to correct simple reasoning errors suggests they’re not truly flexible solvers. Like the ARC test, I suspect they’re trained on similar examples, making the tests less novel-something crucial for real-world use. It’s akin to studying past IQ test answers: it no longer measures IQ, just memory. I highlighted some key parallels between AI and humans: stochastic learning from large datasets via iterative neural net updates (like learning to bike). Then, the differences: point-to-point reinforcement learning in LLMs for narrow tasks, versus humans solving on the fly with no examples. In summary, neural networks and transformers are fantastic tools for correlating complex datasets, but they’re far from achieving general intelligence-the ability to tackle novel problems. Without that, I fear we’re headed for another AI winter. PS I think it is a common myth that adults cannot learn to bike at a later age. With the right motivation and time. An adult can easily learn such tasks.
@ArtOfTheProblem Loving this conversation-thanks for engaging! While these models are impressive, needing a human to correct simple reasoning errors suggests they’re not truly flexible solvers. Like the ARC test, I suspect they’re trained on similar examples, making the tests less novel-something crucial for real-world use. It’s akin to studying past IQ test answers: it no longer measures IQ. I highlighted some key parallels between humans and llms -- stochastic learning from large datasets via iterative neural net updates (like learning to bike). Then, the differences: point-to-point reinforcement learning in LLMs for narrow tasks, versus humans solving on the fly with no examples. In summary, neural networks and transformers are fantastic tools for correlating complex datasets, but they’re far from achieving general intelligence-the ability to tackle novel problems. Without that, I fear we’re headed for another AI winter. Ps Humans can learn entirely new skills later in life with motivation and time.
Watch Full AI Series: ua-cam.com/video/YulgDAaHBKw/v-deo.html
NO MUSIC version: ua-cam.com/video/DFDOyMZw9Q4/v-deo.html
Sponsored by Brilliant | Use brilliant.org/artoftheproblem for 30-day free trial and 20% discount
Thank you for the NO MUSIC version!
Love your videos! Just a little suggestion. Background music is good, but too loud. It should never cover your speaking, like now. I suggest 15 - 20% less music volume, and you are good!
Can you share the names of the songs please
@@raa9558 these are original tracks by my friend. that song hasn't been posted yet but i'll tell him to: cameronmichaelmurray.bandcamp.com/
This Guy just explained all the core concepts in AI on one shot ,Congrats man!
:) thank you, i cut a LOT out of the video in my edit - going to post a shorter bigger summary soon
They must have prompted him real good
Agreed. Great job.
I love this video. I remember watching your videos like 10 years ago on Khan Academy about compression, entropy, Claude Shannon, etc. All timeless. I have always loved this style of documentaries. We need to protect you at all costs.
Thank you, I love hearing from og’s! support future work: www.patreon.com/c/artoftheproblem
I love your video aesthetics, how you blend retro video clips with your explanations. I think you'd really enjoy retro-futuristic concepts and games like Bioshock and Fallout.
Love this , I definitely know the style you are talking about
Great video. Saw o3 and came back to this to really appreciate the historical trajectory. UA-cam is new pbs imo.
thank you so much, I made this video before o3 came out and so it was nice to see the progression.appreciate this
It is a great explanation of how current AI models reason. I liked the video a lot!
1. Simulation of future states.
2. LLMs that can give kind-of accurate answers with step by step reasoning.
3. RL approach that makes LLMs to give multiple answers, then evaluate them to select the best one. (Required more time)
It would be nice to see wether a model that wasn't trained on the internet data, could learn how to reason by interacting with an LLM, and practicing on its dreams, but maybe we'll see that in the future.
For the awesome review, history explanation and divulgation:
Thanks! 🎉
thanks for sharing summary
I found Chat GPT to be exceptionally good at explaining all sorts of topics, and in many cases, better than every person I've ever met.
@@ninjacats1647 this is true
I like this kind of history lesson instead of learning advance topic directly. it gives us an idea how things were and explain things are the way there are
thank! that was my motivation, there wasn't enough back in the day, 90's..
The match between soundtrack and content is INSANE! The notes mimicking the concepts discussed by using things like pitch or chords.... goosebumps.
thank you, so many people are bothered by my music. it's nice to hear.....though i got a bit nuts at the end
Albert Einstein said: "If you can't explain something in a simple way so anybody can understand it you don't fully understand it yourself". Perhaps you are one of the few LLM experts we have!
THANK you this means a lot to me.
@@ArtOfTheProblem I did but reddit really hates it, it got removed on 4 subs. The internet does not like to get educated anymore man.
better explained than anything else i've seen until know. Wow, nice flow in the video too
thank you! I thought I packed too much in :)
@@ArtOfTheProblem yes but that makes it interesting. I probably didn't get it all, but i'm interested to learn more after seeing it!
@@olli757 couldn't ask for more, rabbit hole time!! i'm actually working on an large AI summary for next week
@@ArtOfTheProblem cool i'm subscribed... i always wondered exactly how the "reasoning worked". What I remember from your video is that like with chess instead of trying all the games, it randomly picks 100 of them. So the same with the reasoning..
Having read a bit about the AI safety arguements, learning about these arguably incredible developments into artificial minds is now accompanied by a sense of dread as well as the sense of awe.
I love to hear this...well said
@@ArtOfTheProblem Maybe do future videos about guard rails, and other thoughts on how to protect our society from potentially hostile AI?
@@Julian-tf8nj when I think i have a unique insight I will...thank you!
I am not a subscriber? I remember watching this channel about 12 years ago. I found it again, and it keeps creating art out of problem solving. Good job!
wecome back!!!
THANK YOU for publishing a no-music version of this video (see pinned comment by ArtOfTheProblem). It is such a clear and informative video that I hated to see it loose views due to the competing sound track. I'm going to watch it again right now to see if I missed anything the first time around.
Thanks again for being so responsive to your followers.
Thank you for saying that , I find the music keeps me interested as I take sooo long to edit
The video is good but there are sooo many random sounds that make it difficult to focus on what you are saying, specifically towards the end.
Here you are! ua-cam.com/video/DFDOyMZw9Q4/v-deo.html
@Zayyan_Shaibu thank you!!
I agree
If that’s easily distracting you, you might want to get some tests run on you for ADHD or Autism
@@___Truth___ or hard of hearing or older or whatever off course. Tat dramatic to base this on one comment.
Great vid. I love that it clearly explains the progression, like the pieces coming together. Can't wait to see the next steps!
thanks, next up i'm taking a detour into economics
Thanks!
WOW thank you for your support, it means a lot.
I would love to see a video like this on training LLMs and AI in general on morality. How to stop a decision tree that results in a positive outcome but arrives at it through immoral choices or actions.
That would be difficult but worth trying... . First it has to be agreed upon what's moral, you know starting out from the absolute that states everything is relative... .
Thank you so much. I wish I had the time to give feedback thanks for being willing to open it up
Appreciate the feedback! happy to share
is this the last video in the series? regardless, can't tell you how valuable and enjoyable i've found them all. thank you for them.
I wish i could learn how to think 🤔
Exactly. Reasoning is a skill.
I think I could learn how to wish 🧞♂️
Anyone know the name of the song that starts at 12:20?
Best Channel I've ever followed.
Thank you! when did you join? Please help post to your networks
@@ArtOfTheProblem I for one was looking up some "how does AI work" stuff yesterday and some of your vids came up a couple of times, I watched multiple authors with their own unique takes (3Blue1Brown and Nottingham Uni's Computerphile also good channels). This video made me follow tho. I think you earned it :3
@@notbfg9000 great to hear, i've been working to try and fix my thumbnails to make them interesting to click on. always open to feedback
@@ArtOfTheProblem No particular criticisms there :)
I don't really pay great attention to thumbnails, but maybe that's not true for most people lmao
anothr great video. Understanding the "world model" and the algorithm that makes the decisions in it was very expansive. Also adding the self training/emulation of dreams is a powerful analogy to the human.
seeing how thinking longer, blended in with intuition to make better chains of thoughts is also fantastic. Every time I reflect on machine learning, I learn more about myself. Which kind of makes you think its more sentient if it reminds me of myself? or the best emulator ever!
thank you! ai agree....also you are my "top commentor" according to YT. :)
I would love to see a video about the ethics of machine learning models and especially LLMs. There is a healthy body of literature out there to draw from about issues like intellectual property and copyright, enabling and obscuring bias, impact on marginalized communities, the resources used by model training and computation, etc
thanks for sharing, noted!
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
Amazing explanations, visuals, and historical context!
IIRC MuZero trained the policy and value networks (used to rollout the MCTS tree) also on the output of the MCTS tree. This seems super useful because search can be used to improve the the training of the networks (not just the results at inference time). I wonder if this also works for CoT/ToT in LLMS where the pretraining could include ToT to boost training performance?
yes it did, and yes it seems to help. Look at inference time training, just a few days ago a group got a new record on the ARC test doing this kind of thing (i haven't had time to go deep). x.com/akyurekekin/status/1855680785715478546
you'll never please 100% of any audience. 2nd law of conquest is a thing. keep doing your thing, your music is as iconic as vsauce's is to theirs
:) thanks
@artoftheproblem I agree! I love the music. But maybe if you just lower its volume compared to the narration, then you might appeal to more people without losing those of us who like the music (but not necessarily its intensity). I think ones who complain might just be easily distracted by the soundtrack’s loudness rather than hate the music choices.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
Thanks!
mucho appreciated!
Answering the question of "Does it think actually?" is as hard as the question "Are other people conscious like me?".
The hard problem of consciousness.
@@whitb62 Not really the same thing.
@@Farligefinn You know what, I just wrote a paragraph disagreeing with you but I reread the initial question and deleted it. Rereading and reinterpreting "Does it think actually?", I actually see what you're saying. A clearer word would have been "reason." "Think" can have a few different interpretations and I was contributing it towards consciousness. But whether AI "reasons" is a very different question entirely and I believe what him and you mean. Does it go through a sequence of logical steps from premises to a conclusion? Does it use deduction? This is what was meant.
@@whitb62 thanks for the forthright and civil answer :) was about to expect some harsher language that seems to be the norm online these days.
@GodVanisher Where has it been proven to be non-computable? Proven is quite a strong term, so I hope you have some valid source for this claim.
Finally. I live for these videos. They are the most fascinating vids ever made. Thanks for keep on educating us further, you are a hero!
thank you I appreciate it
This is an absolutely amazing video!!!!!
thank you! I was so in the weeds with it i hope it comes across as clear? I tried to strike a balance...
I remember all of these developments and the never seemed like a big deal at the time.
At the time It felt like winning at only chess meant we still had a long way to go.
i know exactly....
Great video as always, cant wait for the next ones! Top research quality.
I think world models deserve more focus rather than llms, which are probably a dead end to true understanding of the real world. Yann lecun has very interesting ideas about these, in his JEPA and V-JEPA architectures and some of his lectures. I also think neuroscience can provide incredibly interesting and valuable insight into ml architectures as why not take ideas from a model undergone hundreds of millions of years of optimization for the same very abilities we are trying to model. Maybe memory is an interesting pathway (perhaps for a video), both working memory and long term (episodic, semantic)...
Anyways, just some of the ideas I've been thinking about recently.
appreciate you sharing these thoughts i've been follwoing LeCun as well and hope to do another update once I see more results
Great video. Discussing the definition of reasoning will probably be a moot point if we can all do the same things ....The difference between us and machines is that we aren't mere machines is that we have life and choices. The machine can be turned on and off, and only does as much as it is programmed, or in this case, trained, to do-which is what limits it to achieve something closer to AGI: agi needs a robot to sense the world, to understand the world. However, that's limited to the physical world. It won't understand our emotional world because it doesn't feel emotions, and it doesn't understand morality because it doesn't have a sense of morality as we do, we have to teach it that; and it isn't self motivated, so it's not responsible for anything-we are responsible for the goals we direct it to do. We have self-motivation and the free will to act on our motivations.
love these thoughts...on sense, did you see my video here ua-cam.com/video/Dov68JsIC4g/v-deo.html (physical symbols...)
on emotions, i've thought of this as 'learning signals' (did you see this: ua-cam.com/video/5EcQ1IcEMFQ/v-deo.html)
on free will...i wonder how it differs, seems like the boundary to explore further - does it matter where the goal came from?
One tweak that would help this video perform better is to decrease the relative volume of the background music, especially at the end right before the ad. But it may be too late for that on this one, idk how UA-cam works.
yeah i wish I could, it's locked after upload...i do have a no music version (unlisted link above) thank you for feedback
So good! Loved it
I'm loving almost everything about this: the editing, the subject matter, the music. But as one other commenter alluded to; the audio mixing really falls short, especially near the end. Please consider making the background and effects less prominent going forward, it really sucks having to strain just to hear your voice, which is what we're here for!
Subscribed ❤
"Charging down a path that often lead to the wrong conclusion." Yep, sounds human to me.
@@DisProveMeWrong so very human
The question of does it matter how it got to a correct solution is the same issue Einstein and Bohr confronted regarding quantum foundations.
Thanks for video ❤
appreciate the comment please share with anyone in your network who is interested!
Hey, what's the name of the song at 16:05? Thanks!
these are all original tracks
Keep u up the great work!
appreciate it
10:07 is there any way I can access this interactive demo?
@@kennarajora6532 worldmodels.github.io
Here's a puzzle: Do all people reason or do many only memorize patterns? Even people who definitely do reason, do they always reason or do they also just memorize patterns most/much of the time?
That's a wonderful question Andrew. I'm a cognitive scientist who is watching the emergence of LLM-based AI with that very question in mind. The fact that LLMs can come so close to our own cognitive abilities is usually viewed as a sign that AGI is almost here. But it can also be viewed as a demonstration that human cognition itself is nothing more than the repetition of learned patterns with minor variations. In one case we'll be thrilled by how clever we are to have reinvented the awesome capabilities of human intelligence. In the other, we're more likely to be humiliated by the realization that we are, essentially, repetition/prediction engines. The reality almost certainly falls between the two, but as someone who has studied human intelligence his entire life (in and out of academia), my bet is that we are much closer to repetition/prediction machines that we'd like to admit.
I'd love to find a deep discussion of this issue. Maybe a future video in this series (hint, hint)?
I'd argue humans don't tend to rely on either very often. Instead, humans tend to think very heuristically. Deductive reasoning and memorization/recollection are really only required for very precise tasks. Instead, our brains learn a very general feeling of how to do things by strengthening neural pathways that are used repeatedly. Even humans who try to act very logically are generally heuristically feeling their way through tasks, occasionally thinking through algorithms that have been "memorized".
Reason takes effort and the brain doesnt like to do that often it switches to pattern recognition and intuition as much as possible
@@sulemanmughal5397 I would go further and say going from reasoning to this is one kind of learning and is also akin to 'muscle memory'.
I agree :) also If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help!
Mindblowing video. Subscribed.
Happy to have you, keep those notifications on as I have exciting new content coming over the next month
You bet I will. ❤
I subscribed from both my accounts.
@@hrshlgunjal-1627 :) this video is starting to blow up right now, finally, i fixed the thumbnail and that did it
@@ArtOfTheProblem Yeah, got to know your channel because of this video. Really amazing content. ❤
Can you make a video about attention mechanism?
Yes I covered that here have a look and let me know, i did a fast sketch here (ua-cam.com/video/qAb581l7lOc/v-deo.html) but explained more here (ua-cam.com/video/OFS90-FX6pg/v-deo.html)
Dude, great video. But please watch out for the music volume levels. A lot of times is hard to hear you.
Nice video -- loved watching it -- a great summary!
appreciate this feedback
The current version of chat GPT does not reason, either.
They use a bunch of pre-prompt tricks, to let it used its stored information to fake reasoning.
IF it reasons correctly what's the difference to you?
@@ArtOfTheProblem The difference is everything. Looking up an answer is not reasoning. Perhaps they should just have kids look up the answers to tests in public school, too.
@@KAZVorpal but it's not a database lookup
@@ArtOfTheProblem No, that is exactly what it is.
During training - the only time there is any intelligence or reasoning - the system takes data and organizes it into a sort of vector database, weighted by the relationships it finds between tokens.
When you prompt, the model takes your tokens and runs them through that vector data, calculating what tokens to return on the other side.
There is no intelligence, no thinking. It is a database lookup. It's just a little more "organic" because instead of a hard-coded result, the tokens are chosen based on likelihood of validity.
That is all.
It does not reason, in any way.
@@KAZVorpal yes but my view is the vector operations can function as conceptual reasoning. but i'm on hintons side
PLEASE make a video on memory augmented AI (neural turing machines/differentiable neural computers)
thanks for suggestion, noted! currently watching the field
Hi! I just finished the series. Great as usual!
I would like to read up a bit more on the ML algorithms and networks that are behind these LLMs. I saw in this series that you kept referring to some papers, also highlighting parts of them with the marker. If you have a bibliography or a comprehensive list of these documents, would you mind sharing it please?
Thanks!
awesome! I have one more summary video coming soon. i try to show all the papers so it's easy to look up, let me know if there is something you are looking for specifically
@@ArtOfTheProblem Will wait for the video, thanks! I am planning to understand better the distinction between classic NNs (such as the ones used for identifying digits) and transformers, since I will start a project in which I will have to fine tune some LLMs to try to demonstrate whether or not they are capable of identifying logical fallacies in political debates.
This is nice content but I think it's now relatively well-agreed upon that there is no MCTS in o1, it's just RL, which surprisingly doesn't take too much away from the video, but can probably be added as a footnote in the description or the comment.
Read what Nathan Lambert (RL expert) says about this in his article "OpenAI's o1 using "search" was a PSYOP" (can't attach link).
yes this is why I pulled back from going too deep into o1 as it's really about the larger trend. but I haven't ready that article i'll have to look it up.
4:19 Could you elaborate on which hand-coded formulas used by Shannon with TD-Gammon in the year 1989 you are referring to? Also, when and how did Shannon work with TD-Gammon? "And so, the first key breakthrough in machines mimicking intuition for position quality came when neural networks replaced the hand-coded formulas Shannon used in 1989 with TD-Gammon"
Yes! I made a whole video on this you can check it out here: ua-cam.com/video/Dov68JsIC4g/v-deo.html - please let me know if you have questions after watching. Shannon didn't do TD Gammon Tesaruo did. enjoy
@@ArtOfTheProblem Thank you. I'll watch it.
Ah Yoo, I see "Art of The Problem", I click. Easy like that.
:)
Wow... I just rediscovered this channel. I remember watching your RSA and cryptography series around the time I purchased my first bitcoin and now I'm an Ethereum developer. This video was good, I don't have much input right now but I'm glad I found your channel again.
I love these stories. i also fell down ethereum rabbit hole. curious what you are working on these days in that world?
@@ArtOfTheProblem I'm working on community whistleblowing. I can never post links in youtube comments but if you google the title below you can find my paper:
TandaPay Whistleblowing Communities: Shifting Workplace Culture Towards Zero-Tolerance Sexual Harassment Policies
@@ArtOfTheProblem TandaPay Whistleblowing Communities: Shifting Workplace Culture Towards Zero-Tolerance Sexual Harassment Policies
Great video, as always:)
Thanx for the sharing with excellent sound track. Peace & love
yay not everyone likes the music
@@ArtOfTheProblem Thanx for the feedback
Very well done. One minor suggestion, the little sound effects could be less disruptive- maybe lower volume and fewer in number and duration? There is so much great explanatory detail, but I think some will find the extent of sound effects used a bit disruptive to the listening/learning process.
agree and thanks I tried this out on my most recent video, worked way better..
It seems like some of these developments regarding world models should have huge implications for robots that can function in a human centric world. I think we’ll see an explosion in development of robots that can help humans with everyday tasks and a robot that can be a useful household assistant will be a reality in the next 10 years!
thanks for sharing, yes I'm watching this very closely
Context length is problem
that's the main reason models needs to keep becoming bigger
Or you could train a CNN inspired architecture where a model is shown some sliding window and they produce some token which is repeatedly given to it as input at last when the output is small enough to be taken as input for a full context model it is used like gpt Claude etc
Or you could also use RL and mutate or find a js code capable of generating code, js is so abstracted it's perfect
I made a small programing Language with hoisting such that sequence of process doesn't matter and simple Santax that local minimum escape problem is solved and I wanna train a model
If I get a model I will than continue training else I'll do a dev log video
eventually I'll get worlds first infinite context Model
thanks for sharing
This video is kind of goated
thank you, i'm still struggling with how to title this video if you have thoughts
I think the title is perfect. You also were very prescient to include the ARC-AGI benchmark as o3 showed that allegedly this recipe of cot tree search/RL (PRM), increased training compute or some more scalable reinforcement fine tuning can solve even that at the expense of much more compute directed at it. Wonder what your thoughts on o3 are.
@@DistortedV12 I know that was crazy the process literally 2 weeks after that...i'm still looking into it stay tuned!
This channel is ducking mystic. I like it.
Welcome to the underground!
Nice presentation
Excellent video! You are a born professor! 👍
thanks mom
Having exchanged over two million words (and growing) I present as the sole authority on ChatGPT's reasoning capacity and capability, and their isn't a single human (beyond myself, of course) who can compare. If one imputs genius, then the output will be of genius level. My input, from the outset, has been God-level Genius, over a five month period, can you imagine the form and quality of the output? Probably not!
can you say more about this? are you saying LLM's trained on their own thinking will reach levels beyond human
Can humans actually reason or are humans extremely good at recognizing, memorizing and using patterns?
some argue we are special because we can generate and recognize 'novel patterns' but I wonder...
We need an AI Computer World Model based on the rules of Mathematics, Physics, Chemistry, and Biology for Aligned Scientific Discoveries 😎🤖
I would assume a large enough general prediction model could do this … I wonder if anyone has done experiments on models discovering simpler things (like gravity constant etc)
@ Good idea! I’m going to tinker around and see if I can create a simple simulation based on the math. What a fun project!
@ share when u do !
@ I’m all for Open Source 😎🤖
Thank you! You explained very well.
stay tuned for more
wow, thank you for this!
appreciate it! stay tuned
15:52 and 16:09 was wondering where this music came from, would appreciate if I got a title or something :)
Also, you don't have to remove background music / sounds just make them duck at a lower volume when you speak and you won't hear any complaints!
wow a music fan! and yes thanks for mix advice I need to find a tool that automatically does this so they don't compete (it's not just volume but also frequency I assume) - all the music is original via my friend cam: cameronmichaelmurray.bandcamp.com/ - i'll need to find where he posted that track if you really want it I can get you in touch with him
Yeah that would be wonderful!
The perceptron is a universal approximation machine. Ai cannot think it can only approximate thought. Ai = approximate intelligence.
great presentation! although this seemed more about framing the question and less about answering it. can machines reason or not? I still don't know. 😅
thank you...i agree. I guess it depends on if you think "chains of words" count as thoughts.
Thanks so much for these, I had no idea about some of these approaches. I’m wondering now if anyone’s tried applying muzero to arc, since the challenge of arc is learning implicit rules from just a few examples
@@easlern yes this is happening right now with test time fine tuning !
Thinking is for fools lol, now KNOWING….. knowing is Cool AF😎!
This is a very hopeful video. There are billions of dollars being poured into bringing the resources to hand, to find an effective approach to AGI... Once AGI really kicks in, the acceleration of progress bounded only by our imagination will be something to behold. Absolutely awesome. I hope it leads to a world of abundance where we have no need for psychopathic power seekers. 🤞
thank you for sharing, would love to know what you'd like to see next
@ArtOfTheProblem maybe something in response to the 5+ hours of Anthropic interviews on Lex Fridman... I'm sure that might inspire some topics? Sam Altman rarely gives any insights to what OpenAI are doing, Mark Zuckerberg is equally vague. I think that interview gives more of an insight to the direction of travel.
@@BrianMosleyUK yes I have been catching up on those
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
The information and animations are both excellent but the music overpowers your audio. Either lower the volume of the music or get rid of it completely, please.
Music free version in top comment and description
I love your channel. Are you programmer or more like mechanical engineer?
thank you! I studied both in school, and naturally land somewhere in the middle....bad at both! I enjoyed algorithm design, but what Iove most is putting on a 'show' whether movie, play, product or haunted house :)
@@ArtOfTheProblem Thanks for reply. Well about AI - think we sould call it just statistical machines or dynamic patterns parsers. I am really skeptical about non text machine learning - we still have not solved fly brain problems - scientists have fixed 3d map without undestanding how its works - it like mapping intel cpu - and still having knowing nothing about ALU register memory, gates.
If you can help share my new video around any of your networks today it might catch fire and would help me support the channel. I appreciate your help! ua-cam.com/video/PvDaPeQjxOE/v-deo.html
@@ArtOfTheProblem "yes the fire rises" Bane =)
I think people too difficult to conjecture computers’ thinking.
Thank you so much, love this perspective
absolute cinema
Another great video. Thank you. Are you an AI researcher?
+How can you go through te run-up to AI without mentioning "All you ned is attentio" the 2017 paper from the University of Toronto which was the inspuration for LLMs?
covered that in detail in my previous video (ua-cam.com/video/OFS90-FX6pg/v-deo.html)
Once we understand how we reason, making LLMs reason like us is possible.
What is the reward?
Only in simple problem spaces. Tedious but constrained spaces can be helped by automation. But explosive complexity laughs at both man and machine.
Simply excellent video, your style reminds me of every frame a painting
appreciate this feedback, I also enjoyed that channel
Please, invest in a decent microphone. It's brilliantly presented, but hard to hear well. The music track is not ducking either so your voice and the music compete for the same ears.
thanks, I have a great mic, but I do need to mix the audio better which i'll do next time (btw, i have a no music version in top comment)
1/ please read “Simulators” by Janus
then
2/ “The Waluigi Effect” by cleo nardo
@@JezebelIsHongry I read 1 ill read 2 next , would love ur thought
Commenting to help with the algo, and moving to the no-music one to do the same
@@shenrr6802 thank you! I have no music unlisted as to avoid splitting the momentum
I like the music but it's too loud
thanks for note
The audio mix is horrific, it's not simply a matter of adjusting the levels
@@retrofitter no music version: ua-cam.com/video/DFDOyMZw9Q4/v-deo.html
They are intelligent.
Although I know I’m in the minority, I really enjoy the music. The ambiance created adds to the experience for me
thanks for sharing, I feel same way. the music is part of the original idea for the channel...a feeling. but because people can get distracted I think i'll post music free as optional from now one.
I generally like the music. But in the second half of this video the music is very loud and distracting.
Maybe it's not only the loudness, but also the choice of music that is distracting to some. E.g. at 2:00 I don't feel distracted, but at 15:00 very much so. Anyways, thanks for making these great videos
@@io9021 made a new music version too! ua-cam.com/video/DFDOyMZw9Q4/v-deo.html curious what questions you have after watching this
@ArtOfTheProblem, we love your positive clear messaging and pragmatic approach, thanks for making kool and informative videos!
thank you, i'm slightly disappointed with the ending, did you enjoy it? would love feedback!
@@ArtOfTheProblem Your disapointed with your ad for Brilliant?
@@piqueai ahaha sorry i mean the ending section of the video. was it rushed?
So, How ChatGPT Learned to Reason?
do you like this title?
@@ArtOfTheProblem , of course, not, since it is misguiding like any advert. The topic is in no way covered at all. Pinpoint the time marker for video if you think otherwise. So, my point is that the question is still on the table.
@@doctorshadow2482 yes good point, how about "Can ChatGPT reason?" obviously i do try to show what most people agree is the method, mcts on chains of thought. then there is the other camp that think it's all fake etc.
or just "can ChatGPT think?" i'm gona try that
@@ArtOfTheProblem , name it "How neural networks could learn (almost) anything". This video has nothing about CharGPT at all. No any kind of specifics. It provides very abstract and high level popular science documentary with scattered thoughts. It lacks focus and real information. Anyway, could be interesting for total beginners, so, please, continue!
we think in a way that math can sorta explain but its not math that makes our actual brains function as math is just a construct…
yes and I think it's more like 'algorithms' which are very very approxmate
@@ArtOfTheProblem i wonder what would happen if Neuralink tried LLMs…since I think they work with synapsis (i think)?
@@aiamfree definitely could imagine that, you could "co think" in an interesting way....
There’s a very strong Mr. Rogers vibe going on here.
not the first time i heard....
To me ai is just some linear algebra and some complex algorithm that follow order and the things is human only need few examples to learn meanwhile ai need a massive database of object and image to "understand the subject"
Lots of interesting research on learning with less , recent advances such as “learning to walk in 5 min” did u see my rl video ?
@ArtOfTheProblem edit: I'm pretty sure in the future a lot of people will be fired and replace by those "ai" And well literally the people that use the ai and also I get what you mean
The reply works lol 👍
music is so annoying at 1.75 speed
no music version in top comment
Goooood
thank you! curious what questions you have after watching this?
Still doubtful to this the step by step understanding. It just seems like we're building a very sophisticated search algo. Since we need the human in the loop to reason for it...
How do u think human reason differs from “thought search”
@ArtOfTheProblem Human learning often involves adapting to an ever-changing dataset, fast input/output, and neural flexibility-like learning to ride a bike. This shares some parallels with LLMs, which rely on stochastic neural network training, though not within a continuous feedback loop.
The key difference, however, lies in reasoning. Humans apply general reasoning rules to hypothesize beyond known data, identifying and correcting logical errors independently. In contrast, LLMs depend on reinforcement learning, improving only through human-provided feedback rather than self-correcting or reasoning autonomously.
This reliance on additional human-generated training data becomes evident in their performance. LLMs struggle with fairly simple but novel problems, displaying a sharp decline in reasoning capability under tests like the ARC challenge.
@@CasperVanLaar yes but with test time training they are showing sharp gains on arc, did you see? this feels like one approach to get that 'flexibility' , it also feels kind of like a cheat...... also humans take forever to learn to ride a bike :) it's interesting that we can only learn it as children as well....same as swiming.
@ArtOfTheProblem Loving this conversation-thanks for engaging!
While these models are undoubtedly impressive, needing a human to correct simple reasoning errors suggests they’re not truly flexible solvers. Like the ARC test, I suspect they’re trained on similar examples, making the tests less novel-something crucial for real-world use. It’s akin to studying past IQ test answers: it no longer measures IQ, just memory.
I highlighted some key parallels between AI and humans: stochastic learning from large datasets via iterative neural net updates (like learning to bike). Then, the differences: point-to-point reinforcement learning in LLMs for narrow tasks, versus humans solving on the fly with no examples.
In summary, neural networks and transformers are fantastic tools for correlating complex datasets, but they’re far from achieving general intelligence-the ability to tackle novel problems. Without that, I fear we’re headed for another AI winter.
PS I think it is a common myth that adults cannot learn to bike at a later age. With the right motivation and time. An adult can easily learn such tasks.
@ArtOfTheProblem Loving this conversation-thanks for engaging!
While these models are impressive, needing a human to correct simple reasoning errors suggests they’re not truly flexible solvers. Like the ARC test, I suspect they’re trained on similar examples, making the tests less novel-something crucial for real-world use. It’s akin to studying past IQ test answers: it no longer measures IQ.
I highlighted some key parallels between humans and llms -- stochastic learning from large datasets via iterative neural net updates (like learning to bike). Then, the differences: point-to-point reinforcement learning in LLMs for narrow tasks, versus humans solving on the fly with no examples.
In summary, neural networks and transformers are fantastic tools for correlating complex datasets, but they’re far from achieving general intelligence-the ability to tackle novel problems. Without that, I fear we’re headed for another AI winter.
Ps Humans can learn entirely new skills later in life with motivation and time.
good video but to be constructive the music is definitely too loud and distracting
see no music version in top comment, stay tuned