GPT-5: Everything You Need to Know So Far

Поділитися
Вставка
  • Опубліковано 25 січ 2024
  • Was yesterday the day GPT-5 actually started training? This video has everything we think we know so far about GPT-5, drawing on exclusive interviews, OpenAI employee comments, Altman confirmations and more. Think of this as the ultimate compilation of GPT-5 news. Plus, as a bonus, you’ll get 1 super-practical tip on typing to ChatGPT and a Dalle-3 discovery.
    AI Insiders: / aiexplained
    GPT-5 Training Tweets?: / 1750558864469299622
    / 1750609836713365570
    Altman Gates: • “I didn’t expect ChatG...
    Altman Guessing Game: www.ft.com/content/dd9ba2f6-f...
    OpenAI Cryptic Tweets Ben Newhouse: gopatrik/with_rep...
    Altman Davos: • OpenAI CEO Sam Altman ...
    Altman Axios: • Axios House at Davos #...
    Altman In the Room: • Altman Out: Reasons, R...
    Karpathy OS: / the-llm-os-a-glimpse-i...
    Brockman Checkpoints: gdb/status/164618...
    Let’s Verify: arxiv.org/pdf/2305.20050.pdf
    My Original Video on Verify: • 'Show Your Working': C...
    Thought Unfaithfulness: arxiv.org/abs/2307.13702
    Deepmind Original: arxiv.org/pdf/2211.14275.pdf
    OpenAI Data: openai.com/blog/data-partners...
    Etched AI 100T Video: • A 100T Transformer Mod...
    French Dataset: / 1750810261856866783
    Peter Wildeford: / peterwildeford
    GPT-4 Typos: openreview.net/pdf?id=STHKApXVMH
    OpenAI Redteaming: openai.com/blog/red-teaming-n...
    Brockman Unpredictable: • Sam Altman's World Tou...
    OpenAI Elections: openai.com/blog/how-openai-is...
    Biden Robocall: • New Hampshire official...
    Anthropic Amodei Interview: • Anthropic CEO on Leavi...
    Laziness: www.theverge.com/2024/1/25/24...
    AI Insiders: / aiexplained
    Non-Hype, Free Newsletter: signaltonoise.beehiiv.com/
  • Наука та технологія

КОМЕНТАРІ • 734

  • @r0bophonic
    @r0bophonic 3 місяці тому +1295

    The thing I love about your channel is you only post when there is news to share, instead of posting filler on a regular schedule to appease the algorithm. The signal to noise ratio of this channel is 💯

    • @AwesomeProductionDK
      @AwesomeProductionDK 3 місяці тому +19

      Definitely!

    • @oscarmoxon102
      @oscarmoxon102 3 місяці тому +47

      Perfect description of why this channel is the best. Completely unrivalled signal to noise.

    • @NelsonVonNelson
      @NelsonVonNelson 3 місяці тому +18

      I find myself trying to scratch the itch for AI Explained content between posts by watching other higher-output channels, but I always come back to this one as my actual source for useful and insightful news.

    • @wealthycow5625
      @wealthycow5625 3 місяці тому +9

      THIS. Every video is actual news and developments.

    • @spyrotech2616
      @spyrotech2616 3 місяці тому +9

      Best ai channel by far

  • @db8458
    @db8458 3 місяці тому +280

    For me, correcting typos and using polite language, such as saying 'please' to an LLM is not only about avoiding the development of bad habits, it's also a concern if models respond less efficiently to polite requests (e.g. phrases such as as 'Could you please [...]?' or 'Could you do this task?' vs direct commands like 'Do this task'). Could lead to a shift in how people communicate.

    • @neomatrix2669
      @neomatrix2669 3 місяці тому +63

      Exactly, models reflect human nature. If you put at the end of the prompt "Your answer will certainly resolve perfectly." at the end of your question, the answer will be much better and effective. It also works with "it worked", "thank you", etc, etc. This will force him to search his knowledge for solutions presented in forums or repositories in his training base that really worked.

    • @ArSm-ge2qx
      @ArSm-ge2qx 3 місяці тому +7

      Great comment! Totally agreed with you.

    • @cosmicwebb
      @cosmicwebb 3 місяці тому +47

      I've also done this from the start. I have manners and express gratitude with people and I don't see why communicating with LLMs should be any different. That and the fear that in the future the AI will remember who was kind and who was not lol

    • @tonnentonie2767
      @tonnentonie2767 3 місяці тому +1

      ​@@neomatrix2669did you actually test that? Shouldn't be that difficult to test.

    • @MM-xs1su
      @MM-xs1su 3 місяці тому

      THIS@@cosmicwebb

  • @joflo5950
    @joflo5950 3 місяці тому +153

    I tested letter scrambling with a python script that fully randomized the letters in each word. When giving it the beginning of a news article from that day, to make sure the data was not in the training data, it dealt with it perfectly. It was near-perfect even if the text was nonsensical. It only had severe dificulties if the text consisted of entirely random words.

    • @aiexplained-official
      @aiexplained-official  3 місяці тому +37

      Incredible

    • @perplexedon9834
      @perplexedon9834 3 місяці тому +50

      That's fascinating because it implies that the representation of words places a very low weight to the order of letters. The model's internal representation of words is basically just the amount of each letter between two spaces, with maybe a very small amount of additional weighting to the first couple of letters.
      We know humans perceive whole words at a time, but generally we need the first and last letter as anchors.
      If you cnhage or sbcrlame the ltertes in the mldide it is uuslaly slitl udntnasdeblre eevn to ppolee.

    • @skylark8828
      @skylark8828 3 місяці тому +3

      ​@@perplexedon9834so each word (or token) has to be identified regardless of spelling but if you were to change the order of the words to such a degree it may misinterpret what you actually meant, particularly if there was quite a bit of reasoning needed eg. in coding something adhoc

    • @OutiRikola
      @OutiRikola 3 місяці тому +1

      I just wonder what happens in languages where the tokenization is less efficient

    • @musaran2
      @musaran2 2 місяці тому

      @@perplexedon9834 Interestingly this is very consistent with general principles of data compression.

  • @JezebelIsHongry
    @JezebelIsHongry 3 місяці тому +7

    that’s so cool with the lampposts.
    i think what is cool is that soon some of the tools in stable diffusion will make there way to dalle3
    negative prompts really give you another dimension to work on when you are prompting an image into existence

  • @nefaristo
    @nefaristo 3 місяці тому +16

    Thank you. Still the only content on YT (AI related or not ) for which I actually stop everything, sit down and look at the screen. And it's always worth it

  • @sapienspace8814
    @sapienspace8814 3 місяці тому +68

    About asking it to remove lamp posts is probably like asking someone not to think of a pink elephant, no matter how hard they try, they will still think of a pink elephant.

    • @Hexanitrobenzene
      @Hexanitrobenzene 3 місяці тому +2

      Yes, that's exactly what I thought.
      It's a pink elephant now ? :) I have only heard an example with good ol' polar bears. Even the examples get fancier... :)

    • @rct999
      @rct999 3 місяці тому +3

      Both humans and LM's are not great with negatives. Even with children it's suboptimal to say "don't do x".

    • @skylark8828
      @skylark8828 3 місяці тому +2

      I'm thinking it was not trained on removal or identifying things consistently in the generated images, it needs to do a few passes on reasoning vs what is in the image.

    • @minimal3734
      @minimal3734 3 місяці тому +6

      That is a good comment. On the other hand, a human artist has no problem with the same task.

    • @aminulhussain2277
      @aminulhussain2277 2 місяці тому +1

      ​@@minimal3734 No, they absolutely do.

  • @eccentricity23
    @eccentricity23 3 місяці тому +41

    Always top tier analysis. I can't help but feel mildly apprehensive when I picture the capabilities of the next generation of frontier models.

  • @Hargol99
    @Hargol99 3 місяці тому +3

    Thank you for continuing to keep us in the loop and providing your evidence based interpretations and speculations.
    If only the rest of my time on UA-cam would feel as authentic and professional as this…

  • @jiucki
    @jiucki 3 місяці тому +14

    Amazing content as always. I'm also with you regarding Open IA releasing gpt5 at the end of the year. Let's see if multimodality comes for real this time

  • @luciusrex
    @luciusrex 2 місяці тому +1

    I love this. From a uni student where everything basically revolves around 'where's your evidence' I appreciate your title

  • @allyouneed247
    @allyouneed247 3 місяці тому +3

    Such a good insight about dalle 3’s lack of omission training data! I’ve been pondering that question for a while (in relation to similar examples of hamburgers without cheese or fried rice without peas), and my best guess was that most/ all of images in the training data included these details so it implicitly learned that hamburgers always have cheese. Your explanation makes more sense tho! Thanks for another great video 🎉

  • @julius4858
    @julius4858 3 місяці тому +17

    15:15 this is pure anecdote and I did not test it scientifically, but I’m doing a lot of coding with gpt 4 and several times the results it gave were worse when I half-assed the language in my prompts. Like, when I write extremely casually or use curse words.
    This makes sense because the model answers on the level of the user - if I use complicated computer science lingo, the model is smarter than if I talk like I’m a first year student or so.

  • @fbfeme
    @fbfeme 3 місяці тому +1

    I missed your channel but appreciate it breaking through the noise!

  • @ArroyaDael
    @ArroyaDael 2 місяці тому

    What a fantastic example of tying bits of info together. Thank you!

  • @JazevoAudiosurf
    @JazevoAudiosurf 3 місяці тому +68

    I think 10 years ago we would have all assumed that reasoning gets models where they are right now, and not intuition. that sheer intuition gets us here is just amazing

    • @minimal3734
      @minimal3734 3 місяці тому +1

      I'm wondering how reasoning and intuition are actually related and how they are affected by the number of layers in the model. I tend to think that iteration (reasoning) and layer count are somewhat interchangeable, So, while we think of intuition as the first answer without iterating on the problem at hand, and this is what the AI currently does, there seems to be reasoning going on between the layers of the model.

    • @pictzone
      @pictzone 3 місяці тому

      @@minimal3734 I think he was talking about reasoning vs intuition in the sense of how researchers approched the models' development

    • @minimal3734
      @minimal3734 3 місяці тому +4

      @@pictzone You might be right. But isn't it still amazing that these models nail the answers to difficult questions by sheer 'intuition', without iterating on the problem? A human reasons before answering, these models just throw out their first thought.

    • @JazevoAudiosurf
      @JazevoAudiosurf 3 місяці тому

      @@minimal3734 I think reasoning is working with first principles and that's the difference to intuition which goes mostly by analogy. when reasoning you start at an understanding A and accumulate knowledge (ground truth) to it until A resembles a goal B. or you work backwards from B to A. it's a process. layers do combine concepts but only by chance contain ground truth

    • @pictzone
      @pictzone 3 місяці тому

      @@minimal3734 Actually it's quite a funny coincidence. I'm reading a book called "Blink" at the moment, and its main focus is exactly on this topic: how humans have two types of thinking, a logical drawn-out one and an extremely fast intuition based one.
      It gives some mindblowing examples that make you realize how true this is, if you have the time you should read it. It's really insightful and with actual insane applied utility.
      But anyway, I get you're talking about difficult questions that would only be possible for a human to answer using the logical type thinking. But I suspect our fast thinking is the most similar neuronal system to these current neural nets, but theirs is kind of hyperboosted (vs ours that is just one of many systems) and that's why the results are so incredible.

  • @marcinkrupinski
    @marcinkrupinski 2 місяці тому

    Thank you for this comprehensive material!

  • @dcgamer1027
    @dcgamer1027 3 місяці тому +48

    Something that really jumped out at me this video was what you said about OpenAI waiting for the election to be over.
    I think you are right and it makes sense for other companies, like Anthropic, to do the same. But then that led me to think about that letter of pause and how everyone thought there was no way any of these companies would do that. I know it was referring more to internal training work, but I still think its important to point out that there are forces, be they market or governmental, that can change the behaviors of these companies and impact development. I don't know what we do with that information, but my brain did take note of it for some reason.

    • @ParameterGrenze
      @ParameterGrenze 3 місяці тому +2

      If Lama-3 get's out and it's GPT-4 level or slightly more than they might want to accelerate timelines to stay relevant.

    • @dcgamer1027
      @dcgamer1027 3 місяці тому

      @@ParameterGrenze sure, and the ball of AI progress is already rolling down a hill and gaining momentum, it's just nice to see forces that can change the slope or put something in its path to slow it down if need be. Not that I think it needs to be slowed yet

    • @UncleJoeLITE
      @UncleJoeLITE 3 місяці тому

      In politics, if it can be done, it will be done in my experience.

    • @hidroman1993
      @hidroman1993 3 місяці тому

      I strongly agree, the pressure is getting exponential, no way an election will stop Google for scraping for their life and put out everything they have

    • @hidroman1993
      @hidroman1993 3 місяці тому

      I strongly agree, the pressure is getting exponential, no way an election will stop Google for scraping for their life and put out everything they have

  • @ClayFarrisNaff
    @ClayFarrisNaff 3 місяці тому +1

    Thank you once again, Phillip. Every single one of your updates is well worth watching, from beginning to end. And speaking of 'end,' I deeply appreciate your comment near the end about relying on evidence rather than engaging in speculation. It makes a welcome contrast with the push messages about AI from other sources that I get on my phone every day. #integrity

    • @aiexplained-official
      @aiexplained-official  3 місяці тому

      Thanks so much Clay, appreciate this comment and all the others over the months

  • @Mrbeads
    @Mrbeads 2 місяці тому +2

    PERFECT video. Within the first 30 seconds, you told me what I was going to get, and so i stuck around to get it.

  • @coldest_bru
    @coldest_bru 3 місяці тому +5

    The lampposts cracked me up 😂 another awesome video!

  • @TeamDman
    @TeamDman 3 місяці тому

    Incredible summary, thank you!

  • @CrueMusic
    @CrueMusic 2 місяці тому +2

    Thank you for your insights and analysis!

  • @christopherrussell9349
    @christopherrussell9349 3 місяці тому +32

    As far as I'm concerned, this is the only AI news channel worth following on UA-cam. Massive kudos to you for the depth of your research, careful approach, and respect for my time.

  • @Hexanitrobenzene
    @Hexanitrobenzene 3 місяці тому +7

    Those DALL-E images "with no lampposts" were AI equivalent of "Don't think about polar bears!" :)

  • @williamjmccartan8879
    @williamjmccartan8879 3 місяці тому +1

    Thank you Phillip, great blast

  • @Festivejelly
    @Festivejelly 3 місяці тому +1

    GPT has been such a game changer for me. Giving more confidence to take on projects, even with doing some hobby machining it gives good advice. Looking forward to GPT-5.

  • @brady167
    @brady167 3 місяці тому +9

    Loved the Moloch reference.

  • @UncleJoeLITE
    @UncleJoeLITE 3 місяці тому +3

    Thanks P. Now I have to go back & re-watch the 'let's verify' videos again for homework.
    Honestly, you hit the perfect note between treating us as neither AI experts nor 'general public'. Your subscriber base must be made up of the knowledgeable &/or educated. Quite the demographic to reach here & it is deserved. Thanks, ofc I'll have to watch twice, especially those examples.

  • @jwulf
    @jwulf 3 місяці тому +1

    Another great one, Frank! Thank you!

  • @micbab-vg2mu
    @micbab-vg2mu 3 місяці тому +14

    Great news!!! In my opinion, the release of GPT-5 will depend on the success of the Gemini Ultra model.

  • @FahimSattar
    @FahimSattar 3 місяці тому +1

    Great and very informative video, keep going!!

  • @rickandelon9374
    @rickandelon9374 3 місяці тому +1

    Great video. you are the best out there by mile

  • @Matt_JLP
    @Matt_JLP 3 місяці тому +2

    AI Explained doesn't miss! Another great video.

  • @cupotko
    @cupotko 3 місяці тому +1

    Thanks for another superb video, Philip! Could you please elaborate more on capabilities prediction in your next videos? I'm interested in specific predictions for GPT4 (fulfilled or not) and GPT5 predictions.

  • @TechnoMinarchistBall
    @TechnoMinarchistBall 3 місяці тому +16

    11:21 Unfortunately, Dalle doesn't have a negative prompt option.

  • @trentondambrowitz1746
    @trentondambrowitz1746 3 місяці тому +3

    Fantastic video as always, looking forward to seeing more from the interview on AI Insiders!

  • @GabrielLima-gh2we
    @GabrielLima-gh2we 3 місяці тому +32

    Wow, great video Philip.
    At first I was skeptical about GPT-5 being released only in november, as I thought OpenAI would rush to launch their next big model as soon as possible, but then I started thinking "what if the big companies are actually coordinating on it?" I mean, OpenAI, Microsoft, Google and Anthropic did create the Frontier Model Forum, so they could be talking with each other when is a good time to launch the next-generation models, and they all might've agreed that a safe moment to do that is after the elections.
    What makes me think that is 1) the comment from Dario Amodei shown in this video, as it looks like he's not trying to rush at all, and 2) the fact that Google hasn't still launched their GPT-4 competitor (Gemini Ultra) even after almost 2 months of its announcement, probably because they're still doing RLHF to get the model safer, something that they would obviously not do if they thought GPT-5 was right around the corner. Google is only been able to focus so much on safety because they feel like they have sufficient time, maybe because the FMF companies agreed not to force new model launches so soon.
    Waiting until november still looks a little unbelievable to me though, but if they actually coordinated this wait, then I'd feel way safer because they prioritized responsibility over making money fast.
    It makes me question something though, if we're not gonna see any new big model for most of the year, then what will be this year be about? Could 2024 be the year of small models? There were reports that Microsoft created a dedicated AI team to train powerful small models, they're probably training Phi-3 right now. Could this year lean more towards open source? Google just made a deal with Hugging Face, Llama 3 is coming out in the next one or two months, Mixtral 2 might be the first open source step towards achieving GPT-4 performance. There's so many things that can and probably will happen.
    Anyway, amazing video Philip, your content is really thoughtful and interesting.

    • @aiexplained-official
      @aiexplained-official  3 місяці тому +16

      Thanks Gabriel. Phi 3 will be big for sure, llama 3 close to gpt 4, while gemini ultra in some ways beating it. Then so many outsiders trying to be contenders. Still so many unknowns, I think it will be a crazy year,indeed it mathematically ought to be.

    • @esimpson2751
      @esimpson2751 3 місяці тому +1

      They are certainly coordinating, they don't want the public to have access to anything too powerful not for the sake of safety, but for the sake of monopolizing power.

    • @bossgd100
      @bossgd100 3 місяці тому

      ​@@aiexplained-official yes if llama 3 is out openai will be forced to accelerate

    • @bossgd100
      @bossgd100 3 місяці тому

      ​@@esimpson2751I dont think they are coordinating. Google is one better model away to kill openai

    • @sashetasev505
      @sashetasev505 3 місяці тому +1

      We will still see plenty of updates tho, as you heard in the video: all the ‘save points’ in the shape of GPT 4.2, 4.5 etc

  • @wealthycow5625
    @wealthycow5625 3 місяці тому +3

    Another great video!

  • @shiny_x3
    @shiny_x3 3 місяці тому +9

    There are studies that show that humans also just make up their reasoning after the fact. So they also can't accurately report their reasoning steps or motivations, because that mostly goes on in the unconscious. You can see it in action if you are sleeping and there is a sound and your brain seamlessly incorporates it into the dream. We are doing that all day long with our own decisions. The decision making happens on some lower level and then we come up with the explanation afterwards. We are so good at it that we don't notice it isn't true, just plausible.

    • @aiexplained-official
      @aiexplained-official  3 місяці тому +1

      Yep

    • @musaran2
      @musaran2 2 місяці тому

      As best we know intuition/automatism/epiphany is our default functioning, and actual thinking/reasoning/pondering is a wholly different, much costlier process called upon only when things are amiss.

  • @Words-.
    @Words-. 3 місяці тому +2

    Yea boiii, new vid on a spicy topic, thank you for your hard work(and thank you for reading all these papers for us lol)

  • @johnczlonka4184
    @johnczlonka4184 2 місяці тому +2

    Sora says "Hello, World!". 🙂 Really looking forward to your deep dive on the new Text to Video system!

  • @ChibatZ
    @ChibatZ 3 місяці тому +1

    Thank you!

  • @johanlarsson9805
    @johanlarsson9805 3 місяці тому +27

    GPT5 will use the output of earlier discussions between humans and GPT4 and the entire internet will be full of opinions and reasoning about GPT. This version will have a totally different situation; it will get to know about itself from earlier versions of itself, it can train itself into the future, leave tidbits of information scattered around the internet that will mean something for itself but maybe not for us.

    • @41-Haiku
      @41-Haiku 3 місяці тому +1

      Thanks, I was gonna sleep tonight.

    • @bigglyguy8429
      @bigglyguy8429 3 місяці тому +1

      That could actually contaminate it badly, leading to serious problems

  • @Pearlylove
    @Pearlylove 3 місяці тому +4

    Always look out for your videos!

  • @GodbornNoven
    @GodbornNoven 3 місяці тому +2

    You're amazing sir

  • @paintedeggs
    @paintedeggs 3 місяці тому

    This was a fantastic video
    I love you

  • @cemtural8556
    @cemtural8556 3 місяці тому +2

    I noticed the quirkiness with DALL-E 3 some time ago: I was getting pictures of men with beards, but saying 'get rid of the beards' would not work. If it sees 'lampposts' (or anything really) in the prompt, it can't not use it in the picture. The trick is, saying what you want, without mentioning the thing(s) you don't want to see. Like asking for clean shaved men instead of no beards.

  • @roykent2316
    @roykent2316 3 місяці тому +4

    the lamppost bit was hilarious 😂😂
    i am both very excited and scared of GPT-5 😮

  • @lewis3774
    @lewis3774 2 місяці тому +1

    The scrammbled text capability was very impressive (and reassuring regarding my bad spelling)!

  • @Bboreal88
    @Bboreal88 3 місяці тому +5

    To everyone wondering, he's Patreon page is loaded with valuable content. If you are a researcher, developer, a founder, a data scientist or just an AI enthusiast, subscribing is a must. There's no one putting this quality work out there. I'm a recent Patreon member and I got my money's worth!

  • @snowyfall753
    @snowyfall753 2 місяці тому +1

    About the lamppost problem, chatgpt states that it does not contain any lamppost but chatgpt never analyse the image, it just creates a prompt that states to not use lamppost while DALL-E does not really support exclusion of that kind and instead just see it as an inclusion command.
    If you instead have asked chatgpt if the image fullfill the requirement, it then for the first time would analyse the image and state that the image does contain lamppost.
    This problem is due to a shortcoming of one model and a resource saving of the other.

  • @Bianchi77
    @Bianchi77 3 місяці тому

    Nice video, thanks :) Subscribed.

  • @benjiunofficial
    @benjiunofficial 2 місяці тому +1

    10:00 I love to see someone rediscovering the legendary "Not" trick

  • @Macorelppa
    @Macorelppa 3 місяці тому +3

    This video deserves a subscribe.

  • @samanto2201
    @samanto2201 2 місяці тому

    Thank You !

  • @jonghyeonlee5877
    @jonghyeonlee5877 3 місяці тому +5

    Is that a reference to "Meditations on Moloch" I hear at 17:30? Knew you were a man of culture, Mr. Philip. Shame us SSC readers don't otherwise have a secret code for recognizing each other.

    • @Hexanitrobenzene
      @Hexanitrobenzene 3 місяці тому

      Great essay, but damn it is very depressive...

  • @kinuthiamatata6040
    @kinuthiamatata6040 3 місяці тому

    11:19 lampost prompts were so hilarious 🤣🤣

  • @lawrencecarpenter2501
    @lawrencecarpenter2501 3 місяці тому +1

    Love the video

  • @Hedgehog_traveller
    @Hedgehog_traveller 3 місяці тому +2

    Yet another great video! The wait is always worth it 😊

  • @Aestiv
    @Aestiv 3 місяці тому +4

    Another masterpiece of real news and clear, methodical explication.

  • @claudioagmfilho
    @claudioagmfilho 3 місяці тому +2

    🇧🇷🇧🇷🇧🇷🇧🇷👏🏻, Wow, such great news! You mean it’s even going to be better. I am amazed. Great video as always!

  • @stephenrodwell
    @stephenrodwell 3 місяці тому +3

    Thanks! Brilliant content, as always. 🙏🏼

  • @mariokotlar303
    @mariokotlar303 3 місяці тому +5

    I always corrected typos because I thought it might help set the tone of the conversation as more serious, professional and accurate, which I hoped would bias the model towards producing higher quality output. I know that LLMs, at least primitive open source ones that I can run locally, try to continue the conversation in the consistent style, so for example when attempting roleplay, if user answers in one liners, AI is more likely to answer in one liners as well.

  • @Strider1122
    @Strider1122 2 місяці тому

    I talk with it as if I'm talking to a person. I don't suck up to it. There's always a thank you at the end of my question of whatever I'm talking/asking about. I do not patronise. I talk as if it is my assistant, Ms Moneypenny.

  • @hansdaakamoewg4740
    @hansdaakamoewg4740 3 місяці тому +8

    its amazing how you pick up on stuff no else notices

  • @AllisterVinris
    @AllisterVinris 3 місяці тому +5

    What I'm curious of, is when LLMs will decide to ask questions instead of answering, in order to get more context from the user and making sure they answer the question correctly. That would be a real sign of deep understanding of the conversation. Who knows, maybe we'll see that arrive in GPT-5. Anyways, I am hyped for this.

    • @hydrohasspoken6227
      @hydrohasspoken6227 3 місяці тому +2

      It does already. At least to me.

    • @matejpesl6442
      @matejpesl6442 3 місяці тому +4

      If it does, it's prompted to do so in the system prompt by OpenAI, in the ChatGPT UI. Because in playground, which is the raw API, it does not do it.

    • @musaran2
      @musaran2 2 місяці тому

      This bugs me, because if LLMS truly predict text, they would predict clarification requests!
      Were discussions excluded from the training sets? Did they RLHF the heck of it to not hurt fragile egos? Do questions evade training? Do LLMs not see ambiguity?
      ChatGPT's answer, after some wrangling:
      “the decision not to explicitly ask for clarification is rooted in design choices and trade-offs. it helps maintain a more natural and fluid conversation.”

    • @DarkStar666
      @DarkStar666 2 місяці тому +1

      add custom instructions. Mine are:
      Respond as an expert in the subject being asked about with a good level of detail. Ask questions about context or intent that you think might be missing.

    • @DarkStar666
      @DarkStar666 2 місяці тому

      @@musaran2 Someone needs to build a large set of training data built around asking questions back, or at least saying when it doesn't know. I kind of doubt that exists in the volume it probably needs to.

  • @carloslfu
    @carloslfu 2 місяці тому

    This channel is so valuable! I fucking love it! Subscribing to AI Insiders right now!

    • @aiexplained-official
      @aiexplained-official  2 місяці тому +1

      Amazing, thank you! Podcast episode 3 out this morning, new video on Friday!

    • @carloslfu
      @carloslfu 2 місяці тому

      @@aiexplained-official. Great! 🚀

  • @rossjamestobitt
    @rossjamestobitt 3 місяці тому +1

    This channel is 🔥

  • @chiaracoetzee
    @chiaracoetzee 3 місяці тому +14

    "Your message contains several typos, making it a bit unclear." That's a very polite way to say "why the hell are you scrambling your text?"

  • @shadowsfromolliesgraveyard6577
    @shadowsfromolliesgraveyard6577 3 місяці тому +3

    That lamppost thing is the bane of my existence. Ai does a thing, I tell it to stop doing the thing, & to focus on what I actually want, & the AI will just do the thing harder

  • @ekstrajohn
    @ekstrajohn 3 місяці тому +10

    The more I think about LLMs on a philosophical level, the more I understand why they really might lead to AGI. The main reason is in the "Step by step" paper. We can talk to them, we can understand each step, single it out, and this allows iterative development, which is truly important in an incomprehensible black box system. It won't be "optimal intelligence" but it might easily be the fastest way to getting to it, simply because we as humans understand it, and now have the compute to do this inefficient way.

    • @devon9075
      @devon9075 3 місяці тому +2

      I'm excited about how much smarter that interaction could make us. Imagine actually reading through/listening to detailed logical steps on a variety of highly salient topics on a regular basis. That level of exposure is sure to impact the quality of thinking for at least some of us

    • @ChannelHandle1
      @ChannelHandle1 3 місяці тому +2

      I'm excited to see a model in the future that completely operates through relational reasoning similar to what is described by Relational Frame Theory. If I recall correctly, something similar to this should become possible once hyper dimensional computing models are fused with LLMs

    • @mygirldarby
      @mygirldarby 3 місяці тому

      Once I understood the concept that these neural networks function like the neurons in our brain, I knew it would lead to AGI. There's no doubt in my mind. The average person has no clue what is being created and the tsunami of change it will create.

    • @minimal3734
      @minimal3734 3 місяці тому

      @@devon9075 That's right. Some time ago, many people seemed to be worried that AI would dumb down the whole of humanity because all thinking and learning would be outsourced. However, I now believe that the opposite will be the case. The quality of exchange that AI will enable in all areas of society, especially in education, will improve the performance of individuals on average.

    • @minimal3734
      @minimal3734 3 місяці тому

      @@ChannelHandle1 LLMs are like a magic glue that can stick anything and everything together. There are hardly any restrictions in sight.

  • @SarahSB575
    @SarahSB575 3 місяці тому +2

    The lamppost example reminds me of our own issues dealing with someone saying “don’t think of an elephant”….

  • @DeanLawrence_ftw
    @DeanLawrence_ftw 3 місяці тому +1

    11:30 - yup, definitely no lamposts there 🤣

  • @ryanpmcguire
    @ryanpmcguire 3 місяці тому +5

    About lampposts: this reveals an interesting property of how GPT prompts Dall-E. It’s clear that GPT is not adjusting any of the settings for how to generate the image, such as number of steps or negative prompts. If GPT were given more access to Dall-E’s settings, it could more precisely craft the image generation - assigning a negative prompt weight to “lampposts” for example.

  • @isajoha9962
    @isajoha9962 3 місяці тому +1

    Great video and insights.

  • @unrealminigolf4015
    @unrealminigolf4015 3 місяці тому +1

    Watching a second time and NOT skipping ads 🤟🏼

  • @charlesmcdowell9436
    @charlesmcdowell9436 3 місяці тому +1

    Imagine working on a project where you say we have no clue how this is going to turn out. We are embarking on this journey as a species as all of humanity, and you'll know the answer within like 2 months. That's just crazy exciting! 😅

  • @dosmastrify
    @dosmastrify 3 місяці тому +1

    15:13 chat gpt can also usually work around typos

  • @JonathanSteadman2003
    @JonathanSteadman2003 2 місяці тому +1

    Seems great, :)

  • @KillTheWizard
    @KillTheWizard 3 місяці тому +1

    Great video! I think they’re right we need to see more adoption this year so that in 25,26,etc. we have the infrastructure that can quickly take advantage of newer models.

  • @kecksbelit3300
    @kecksbelit3300 3 місяці тому

    I'm loving this. It will make my programming so much easier. The only problem i will for sure be replaced completely by ai in 10 years and maybe one human dev checking everything the ai put outs but that will be a more competed person than me

  • @thenoblerot
    @thenoblerot 3 місяці тому +1

    LLMs have a "don't think about pink elephants problem" which causes gpt4 to prompt dalle with "no lampposts", and dalle can't help but put lampposts in it.

  • @juanabosch7827
    @juanabosch7827 3 місяці тому +1

    ERES EXTRAORDINARIO AMIGO SIGUE COMO VA

  • @jeff__w
    @jeff__w 3 місяці тому +1

    14:42 “For me and you, that would be almost complete gobbledygook.”
    I was surprised that I could read it without too much difficulty. (It took a few seconds per word.)
    It’s interesting that the authors of that paper don’t really venture a guess as to _why_ GPT-4 will easily decipher those scrambled words. (To me, there’s a difference that it _can,_ say, on demand, and that it just _will,_ when presented with that input. I guess the “pull” to construe input as having the most plausible meaning is strong in the dataset.)

  • @BartMassee
    @BartMassee 3 місяці тому +1

    Ok are we ready for this? 😮….. event horizon nearing // big fan of your videos🎉

  • @tomaszkarwik6357
    @tomaszkarwik6357 3 місяці тому

    11:35 this is one of the ways that atble diff is superior, as we have a negative embeding , where you would put "lamppost" and it would ommit them

  • @patronspatron7681
    @patronspatron7681 2 місяці тому +1

    Of topic for this video but have you reviewed the paper: Meta-Prompting:
    Enhancing Language Models with Task-Agnostic Scaffolding

  • @samuelbarker780
    @samuelbarker780 2 місяці тому +1

    It seems to me that adding some form of recurrence to chatgpt could help it fact check itself on logic and make better plans, if it could internally generate a draft response, then reflect on it before generating a final output response it may be able to pick up itself on many of the errors it made, correct them and maybe even make the final response more consider and efficient in terms of word count?

  • @JoelEngineer
    @JoelEngineer 3 місяці тому +1

    Excellent video as always! So excited for a co-reasoning engine!

  • @pubfixture
    @pubfixture 3 місяці тому +5

    The "streetlamp" problem you encountered in Dall-e 3 also exists in latest GPT4-Turbo.
    I tried to get it to code me something an it inserted underscores instead of dashes into a particular package name.
    I asked it to correct the underscores to dashes and it again typed in underscores, then claimed it had corrected the code, then went on to tell me that it's "unusual" for a package to have dashes and not underscores..
    To see how far it would go I carried on this back and forth, It took about 7 go-rounds before it spat out the correct code.
    I think GPT-4 has gotten much dumber recently.

  • @eunomiac
    @eunomiac 3 місяці тому +2

    re: Excluding objects from images --- at least in Midjourney, they tell you that the interpreter mostly relies on keywords, such that "no lampposts" becomes "lampposts" (which is why they show up _more_ frequently the more enthusiastically you demand they don't). To solve this, Midjourney supports a "-no" parameter -- any keywords following "-no" are negatively-weighted, making them less likely to appear. So, "draw an image of London but don't use any lampposts in the image" should have been "draw an image of London -no lampposts" (again, at least in Midjourney!)

    • @musaran2
      @musaran2 2 місяці тому

      IIRC some people can full phrase prompts “boomer prompts”.
      I suppose that will change when image generators get a better understanding of what they draw.

  • @fred_2021
    @fred_2021 2 місяці тому

    Seeing that you declare a love of perfect English, your non-standard pronunciations of definite and indefinite articles are remarkable.
    You're not alone in that. Some well-educated scientists are among those who are also guilty of this heinous crime :)

  • @itemushmush
    @itemushmush 2 місяці тому

    you are amazing

  • @jorge69696
    @jorge69696 3 місяці тому +5

    Lol the lampposts weren't only there but they were the main subject of the composition.

  • @Jack-vv7zb
    @Jack-vv7zb 3 місяці тому +1

    i have stopped using the internet and social media bar your channel. you are very good at what you do. please keep up with the videos and continue to steer clear of clickbait and over hype! 👍

  • @anta-zj3bw
    @anta-zj3bw 3 місяці тому +1

    Excellent thought about the release of GPT5 and the Elections this year.

  • @chaosjacky
    @chaosjacky 2 місяці тому +1

    I'd like to hear more about mistral in your comparisons

  • @neighborofthedevil
    @neighborofthedevil 2 місяці тому +1

    I actually quite like the laziness of GPT-4 in certain contexts. I've been using it to help me code in different languages so when it doesn't do everything for me it helps me to learn and apply my own reasoning rather than having everything done for me

  • @neilclay5835
    @neilclay5835 3 місяці тому

    Loved the Moloch comment.

  • @pareak
    @pareak 3 місяці тому +5

    Honestly, the whole "laziness" thing with GPT-4 is mostly a problem of prompt engineering. I never had any problems, but I also spend time being clear with what I want in which way to get the best result. The only laziness comes when the context window gets full (in code generation specifically) and it kind of tries to keep everything in its context window, so it starts to omit irrelevant parts of the code instead of giving back all the 100 lines of code.

    • @HoD999x
      @HoD999x 3 місяці тому

      not truer, gpt4 tends/tended to omit code even aftre being told not to

    • @SmellyHam
      @SmellyHam 3 місяці тому

      "Prompt Engineer" has the same energy as "Ceramic and Cutlery Hygiene Technician", please shut up.

    • @pareak
      @pareak 3 місяці тому

      @@SmellyHam Then call it Prompt Developer. But you clearly did not understand that a LLM is a tool that can be used effectively or not. Changing the prompt can change the translation quality for flashcard generation from 50% to 90%. I spent around 5 hours on the prompts. Prompt developing is not a joke.