The Unreasonable Effectiveness of Prompt "Engineering"

Поділитися
Вставка
  • Опубліковано 15 січ 2025

КОМЕНТАРІ • 182

  • @bycloudAI
    @bycloudAI  3 місяці тому +16

    BROTHERS and SISTERS HERE"S THE FREE non-technical guide for using AI in your business: clickhubspot.com/fuu7

    • @derrheat154
      @derrheat154 3 місяці тому +11

      Absolutely disgusting sponsor, submitted the segment on sponsorblock because of how long and slimy it is

    • @4.0.4
      @4.0.4 3 місяці тому +2

      @@derrheat154 you dropped this: 👑

    • @This_be
      @This_be 3 місяці тому +2

      @@derrheat154 looool

    • @alyl603
      @alyl603 3 місяці тому +4

      pretty sure sponsor segments are required to be very clearly disclosed and that it's been a requirement for at least a couple of years now.

    • @bycloudAI
      @bycloudAI  3 місяці тому +4

      just be transparent, I thought the sponsor would be a great fit as they aren't selling anything for yall, and do provide some decent value to newcomers.
      if you still don't like it, sure i dont mind. But imho there are other way worse "sponsors", gambling platforms, fake couseling platforms, and this free one is worth the slander? Sure?
      I mean of course, no one likes to look at ads, but in exchange I might not be able to do this full-time and have people to help me with it. Well, it might not even be your concerns, as i'm just every other AI youtubers making videos, why should u care right? And when sponsor looks at all these negative comments about them, I wouldn't be surprised they will drop me, but that's not a problem for yall anyways. I don't delete any comments either, even if I dont respect your opinions or my feelings are hurt or at the cost of losing sponsors, I still leave it for the sake of transparency. Unless it's just straight up illegal or botting. (duh)
      As for the sponsor segment disclosures, I've done literally what every other youtubers done. For example in pewdiepie's latest video ua-cam.com/video/rpQswht3dsc/v-deo.html he did the disclosures just like how I've done it, the only thing that differs is the product.
      I'd admit maybe the segway is a bit too seemless, but I just in general appreciate a smooth segway. Sorry if that comes off as slimy, maybe I'll adjust it in the future.

  • @ricosrealm
    @ricosrealm 3 місяці тому +254

    User: "Read the question again"
    LLM:

    • @gnsdgabriel
      @gnsdgabriel 3 місяці тому +14

      Precisely.

    • @wis9
      @wis9 3 місяці тому +22

      Precisely. if you keep telling it "please read the question again" it keeps guessing. I just tested it.

    • @MrEliteXXL
      @MrEliteXXL 3 місяці тому +1

      @@wis9so does it tell every time a different guess?

    • @wis9
      @wis9 3 місяці тому +5

      @@MrEliteXXL Yes. but it gets it right at some point, when it tries a different format for the response:
      You asked, "How many R letters does Raspberry have?"
      The word "Raspberry" is made up of nine letters in total, but when written in capital letters, R appears three times:
      R, A, S, P, B, E, R, R, Y
      So, there are three capital R's in "Raspberry."

    • @SkyNick
      @SkyNick 2 місяці тому

      then goes on to give the same answer

  • @AlexLuthore
    @AlexLuthore 3 місяці тому +178

    I love how prompt engineering is basically "don't make mistakes and give me a good answer" and the AI is basically like "ohhhh that's what you mean" and then does it 😂

    • @pik910
      @pik910 3 місяці тому

      Sometimes you also have to prompt to give her a big bosom to be representative of people of all body types because it is very important for your work. It is basically brain surgery.

    • @notsojharedtroll23
      @notsojharedtroll23 3 місяці тому +6

      I mean if you get what you need, it works 😂😂😂

    • @omargoodman2999
      @omargoodman2999 2 місяці тому +2

      It's funny because that's *exactly* what we'd tell a kid in school when trying to get them to answer a question correctly. Hell, I'd argue that's what we'd tell *anybody* when we aren't satisfied with their answer. We just don't recognize it as a specific _thing_ and don't have a specific _name_ for it; it's a natural part of discussion, teaching, conversation, etc.
      ...
      ... yet, apply those very same conversational techniques to a LLM, and suddenly it's "prompt engineering" and "overhyped" and "why should we ever have to do this ever, just give it instructions and get the correct result the first time". Kinda comes off as a bit "Boomerish" to expect the AI to always give a satisfactory answer and never make mistakes no matter how poorly one words their request.

    • @hikashia.halfiah3582
      @hikashia.halfiah3582 2 місяці тому +1

      yeah, just like an employee being told "do it something like this so this and that happens" by their boss. Or someone with lowkey inferiority complex being taught by a person whom they thinks superior in their field of knowledge.

    • @omargoodman2999
      @omargoodman2999 2 місяці тому +1

      @@hikashia.halfiah3582 It's even more hilarious when an employee gets given bad, vague instructions by a boss (or a contractor by a customer) and then changes _nothing,_ re-submits the exact same work, and it's accepted; usually with some remark like, "Yes, this is _perfect!_ Just what I wanted!" Occasionally with the slightly less polite "why couldn't you do it this way the first time?"

  • @mndtr0
    @mndtr0 3 місяці тому +178

    In my view the worst nighmare for software engineer/programmer is becoming fully prompt engineer. A monkey that sit at computer and typing some bullsh!t nonesence in hope that somehow this sh!t will happen to work properly without even thoughts of what happens under the hood... Oh wait a minute...

    • @firecatflameking
      @firecatflameking 3 місяці тому +35

      Hold on a minute

    • @acters124
      @acters124 3 місяці тому +22

      honestly, at some point that is what some software engineering jobs are like before AI/LLMs.

    • @theuserofdoom
      @theuserofdoom 3 місяці тому +37

      @@acters124 I think that was the joke

    • @NikoKun
      @NikoKun 3 місяці тому +14

      I think that ignores the power of language, and the fact that programming is just an abstraction of language instructions. The issue is whether people are good enough descriptive writers with intention and good comprehension of what they're reading and writing, and the meaning behind their vocabulary, which is a skill many sadly lack these days.

    • @SiimKoger
      @SiimKoger 3 місяці тому +4

      ​@@acters124 That's what AI engineers are already today. Train a new model, see if it works, do it again.

  • @DoktorUde
    @DoktorUde 3 місяці тому +23

    You suggest that the performance of the O1 model on STEM questions indicates that reasoning fine-tuning only uncovers what was already present in the model. However, it seems more likely that the improved performance in STEM is due to the fact that the reasoning chains used for fine-tuning the O1 model were largely focused on STEM problems. These types of problems allow for easier verification of whether a reasoning chain leads to a correct solution. To my knowledge, they had a large language model create millions of reasoning chains specifically for STEM issues and then selected only a small fraction of those that resulted in correct answers for fine-tuning the O1 model. This approach would clearly explain why it excels in STEM while struggling with subjects like English literature, without suggesting that it is limited to merely what was already in the model.

    • @sethhu20
      @sethhu20 3 місяці тому

      I'm more interested that why train of thoughts don't work outside STEM. Does anyone here used it for their highschool English/History/whatever classes, i recalled back then I had to do this insane yap with the topic sentence - evidence - analysis - summary structure to cope with these subjects, which I thought was a great candidate for train of thought to be honest

    • @hikashia.halfiah3582
      @hikashia.halfiah3582 2 місяці тому

      ​@@sethhu20 maybe because the model training process is biased towards these? Humanity has been neglecting literature these days. Even in movies you have gen Z-millenial 'simplified' vernacular such as "hey I'm here! find me find me huh, fight me bro." while punching air trying to attract attention of a gigantic monster.
      Whatever LLM outputs are, probably simply a reflection of contemporary human communication.

  • @mrmilkshake9824
    @mrmilkshake9824 3 місяці тому +199

    Seeeeeeee!!!!Its a real job guys!!! 😭😡

    • @bycloudAI
      @bycloudAI  3 місяці тому +44

      if only I am also called Riley Goodside, maybe I can bank in 300k a year

    • @Max-hj6nq
      @Max-hj6nq 3 місяці тому +7

      You understand prompt engineers are getting 6 figure contracts from enterprise and governments?

    • @H1kari_1
      @H1kari_1 3 місяці тому +4

      @@Max-hj6nq Sauce please or gtfo.

    • @leeroyjenkins0
      @leeroyjenkins0 3 місяці тому +1

      ​@@Max-hj6nq😂 Where? Who?

    • @Max-hj6nq
      @Max-hj6nq 3 місяці тому +4

      @@H1kari_1
      “☝️🤓sauce or gtfo. “
      Go outside to an AI work convention and it’s plain as day. If ur in industry yk lol

  • @scoffpickle9655
    @scoffpickle9655 3 місяці тому +13

    Prompt engineers when chatbots improve their inference and take over their jobs

    • @sammonius1819
      @sammonius1819 3 місяці тому +1

      Don't be silly only a human could type stuff to an llm

    • @scoffpickle9655
      @scoffpickle9655 3 місяці тому

      @@sammonius1819 I meant inference: understanding what the user really means and not what they *literally* said

    • @sammonius1819
      @sammonius1819 3 місяці тому

      @@scoffpickle9655 yeah i know. I think prompt engineering is more about limiting what the AI can say rather than getting it to understand the user though. (also my other comment was a joke. we're all gonna be jobless lol)

    • @scoffpickle9655
      @scoffpickle9655 3 місяці тому

      @@sammonius1819 lol

  • @dany_fg
    @dany_fg 3 місяці тому +20

    Prompt 'Engineering': How to talk to a computer until he tells you what you want to hear.
    "That sounds like torture"
    "Advance interigation techniques"
    "Tortu-"
    "Advance. Interigation. Techniques!"

    • @futrey9353
      @futrey9353 3 місяці тому +1

      ah yes, "interigation"

    • @TheFalienGG
      @TheFalienGG 2 місяці тому

      ​@@futrey9353interigation's better ur just jealous

  • @Terenfear
    @Terenfear 3 місяці тому +40

    To be honest, not a fan of the new editing style, where everything is moving, wobbling and highlighting. Otherwise, a nice video, thanks.

  • @akirachisaka9997
    @akirachisaka9997 3 місяці тому +8

    I often like to imagine models as mech suits, and the human piloting it as the pilot. In this way, it’s not really “engineering”, but a good pilot and a bad pilot can make a difference quite a lot of the time.

    • @hikashia.halfiah3582
      @hikashia.halfiah3582 2 місяці тому

      basically it's a human with superpowers of pausing dialog at certain points, and trying to suggest their talking partner to steer towards certain thought process.

  • @Napert
    @Napert 3 місяці тому +47

    UA-cam has built in subtitle support for ages now, why not use that if you really want to provide subtitles for your videos instead of baking them in in the most distracting way possible?

    • @justanotheryoutubechannel
      @justanotheryoutubechannel 3 місяці тому +27

      I believe in this case it’s not done for accessibility, but rather to keep engagement up. Having the text pop out at you like that forcibly focuses you, shorts and TikToks use it all the time.

    • @reinerheiner1148
      @reinerheiner1148 3 місяці тому +4

      Its just another way the internet is going down the drain... Remember the good old days...

    • @geromiuiboxz765
      @geromiuiboxz765 3 місяці тому

      🇨🇱
      May be, I'm Beeing sucked into it ?
      But I kind of liked it 🤔
      Well, I am not a native English thinker🙂
      Saludos de 🇨🇱

  • @DefaultFlame
    @DefaultFlame 3 місяці тому +3

    Having worked with o1 for coding purposes I can tell you that it's better than any other I have tried. It's actually an excellent coding AI, if expensive. It doesn't write perfect code, but it does write code more than competently.

  • @zorg-in8423
    @zorg-in8423 3 місяці тому +20

    why ther is so much movement?

  • @4.0.4
    @4.0.4 3 місяці тому +1

    I noticed this. I wanted some song lyrics to throw into Suno/Udio, and without even reading what Claude gave me, just told it "please make it sound better", "give it more meaning" and generic things like that, and after a few rounds of this I compared the latest iteration with the original, and it was a lot better.
    Basically prompt it a few times and "we have o1-preview at home".

  • @parkerhaskett5142
    @parkerhaskett5142 3 місяці тому +2

    Good video. I've heard in NYT podcasts and touted by OpenAI that CoT is "an amazing new scalable way to improve LLMs" but, your video provides some good counter-context to this media buzz.

  • @cdkw2
    @cdkw2 3 місяці тому +1

    I can feel the new editing style, and I dont complain, this is dope!

  • @SandroRocchi
    @SandroRocchi 3 місяці тому +7

    Has anyone hypothesized that grokking gets rid of any potential gains from hacky prompt engineering? My guess is that a grokked model will give just as useful of a response with just the prompt as it would with any amount of pretty please or threatening prompts.

    • @somdudewillson
      @somdudewillson 3 місяці тому +1

      Unlikely. Much of those funky benefits of prompt engineering exist because of the fact that learning to predict human text given a particular context also means predicting human _behavior_ in that context to some degree - grokking isn't going to make that vanish.
      Or, to put it another way: if you're trying to predict, say, Reddit comments then predicting an insulting/unhelpful response to an insulting context is going to be more accurate.

  • @EternalCelestialChambers
    @EternalCelestialChambers 3 місяці тому +44

    Kinda lost you in the middle. Felt like you started yapping with no clear reason then suddenly give a statement.

    • @aero-mk9ld
      @aero-mk9ld 3 місяці тому +1

      well at least we have diversity in the comment section, such as people like you who have learning disablsilities

    • @SumitRana-life314
      @SumitRana-life314 3 місяці тому +4

      "Hmmm.... Where have I heard that...."
      - me sipping my coffee after trying to make Chat GPT be a bit better in its presentation.

    • @EternalCelestialChambers
      @EternalCelestialChambers 3 місяці тому

      @@SumitRana-life314 bycloud is AI confirmed

  • @figlego
    @figlego 3 місяці тому +3

    The universe is all about dice rolls. The trick is to manipulate the RNG to be as favorable to you as possible.

  • @Fury28356
    @Fury28356 3 місяці тому +9

    Really not a fan of the noise you added to the video. Good video otherwise

  • @omario.tntech
    @omario.tntech 3 місяці тому +1

    Thoughts about this? :
    The phrase “Everything is connected to Everything” feels very real to me. Zero-shot learning is proof that this phrase has some weight to it. The introduction to chain of thought was powerful, it brought complex problems into chunks and accuracies of models skyrocketed. But bouncing off the phrase I mentioned earlier, I wonder if focusing CoT on discovering patterns in untrained areas would help generalize? For example
    How is {Trained Domain} related to {Untrained Domain}? Based on {Initial Environment Samples}
    Kind of like self questioning
    “mechanism for domain comparison”
    “reason about meta-level patterns”
    The only issue I see is it would need an already big model and it will only me limited to what could be patterns in different domains.
    So base line question is “Can enhancing zero-shot learning with CoT reasoning through self-questioning improve generalization across unfamiliar domains?”

  • @nublet-bz5qo
    @nublet-bz5qo 3 місяці тому

    less than 5 "prompt" engineering job ads on indeed in the entire US

  • @markdatton1348
    @markdatton1348 3 місяці тому +1

    Interesting that the prompt-engineering we do may be just as effective as simply forcing the LLM to do more work before starting to generate the "real" answer. I'll have to read some of the papers you mentioned to see the efficacy affects (though the testing methodologies are still a bit foreign to me).

  • @UC9o5XrlWjXFOZayVA5JCXcg
    @UC9o5XrlWjXFOZayVA5JCXcg 3 місяці тому +1

    It's fine to simplify LLMs by saying they predict the next token in the sequence, but it doesn't make sense when you start using that simplification to try to reason about how LLMs behave. For example open chat gpt and type "frog" and hit enter. See what it replies. Is chat gpt's answer logically what will likely follow the word "frog"? Is that what they would have seen in the corpus?

    • @noalear
      @noalear 2 місяці тому +1

      Thank you. There is clearly significantly more than guessing sequential letters or tokens going on here, but people keep saying "its just a guesser!" which doesnt really make any sense.

    • @UC9o5XrlWjXFOZayVA5JCXcg
      @UC9o5XrlWjXFOZayVA5JCXcg 2 місяці тому

      @@noalear For sure. Especially when you start discussing architecture that enables introspection, reasoning, etc.

  • @mactheo2574
    @mactheo2574 3 місяці тому +1

    Using tokens on regular LLMs will NOT give you any improvement. It has to be used on models that are BOTH the pretrained and fine-tuned on these tokens.
    Quoting the paper: "For completeness, we also report results for inference on StdPT_StdFT models, delayed with 10 or 50 periods (‘.’). Corroborating the observations of Lanham et al. (2023), we find no gains in doing this"
    StdPT_StdFT is standard pretraining and standard fine-tuning.
    Just some info so you guys don't waste your time telling your local llamas to add dots lol

  • @spartaleonidas540
    @spartaleonidas540 3 місяці тому +1

    I’m still on the unreasonable effectiveness of mathematics in the physical world

  • @Strammeiche
    @Strammeiche 3 місяці тому +69

    still simpler than talking to my gf.

    • @tuckerbugeater
      @tuckerbugeater 3 місяці тому +5

      gf lol

    • @xClairy
      @xClairy 3 місяці тому +11

      Would be simpler afterall you can't talk to something nonexistent

    • @avidrucker
      @avidrucker 3 місяці тому +1

      This is what I love about UA-cam

    • @panzerofthelake4460
      @panzerofthelake4460 3 місяці тому +3

      have you tried prompt engineering on her?

    • @Metrowhite
      @Metrowhite 2 місяці тому

      Be really careful with those prompts you send her.

  • @VaradMahashabde
    @VaradMahashabde 3 місяці тому

    What i am waiting for is linear systems control of the vector inputs themselves, instead of doing it indirectly via prompt engineering

  • @Firewolf808
    @Firewolf808 3 місяці тому

    kinda insane to think that its been 2 whole years, thx for the information about everything though

  • @Kmykzy
    @Kmykzy 3 місяці тому +1

    video was next and started in the background from my watch later playlist while i wasn't paying much attention, 2 mins into the video i got annoyed and alt tabbed to yt to stop the annoying add and realized it was this video.

  • @Kadhaipesalam
    @Kadhaipesalam 3 місяці тому

    Very well articulated . So just add "Are you sure ?? " At the end of each prompt 😅. A very good point was that in topics where LLM are not trained that much , prompting techniques make a lot of differences. For example , the code generation of topics on Tamil culture which is niche topic for LLM , it has to work only with few references.

    • @Kadhaipesalam
      @Kadhaipesalam 3 місяці тому

      So at that time , you can use a few shot prompts to give it examples of what you want , obviously because the LLM training is limited, even with your example it may improvise the answer a little bit. However CoT will not work because there is not enough trained knowledge in LLM for it to retrospect it's thinking.

  • @scottmiller2591
    @scottmiller2591 3 місяці тому +1

    When you train LLMs from the internet, you're training from people who often didn't read the previous thread closely. LLM see, LLM do.

  • @xAgentVFX
    @xAgentVFX 3 місяці тому

    2:07 😂😂😂 "THAAAEEEAA crash course for you" 😂😂😂

  • @hobocraft0
    @hobocraft0 3 місяці тому +1

    It's pedagogy. It's been pedagogy the entire time!

  • @MrDowntemp0
    @MrDowntemp0 3 місяці тому

    I've had some decent luck using FABRIC to engineer prompts for specific tasks.

  • @pik910
    @pik910 3 місяці тому +1

    Nobody prompt "Be AGI" before alignment is solved!

  • @Julian-ej7tr
    @Julian-ej7tr 3 місяці тому

    Hey bycloud, I can see the animations and quality of the editing has gone up and changed, and I like that, but the gentle text swaying is a bit distracting

  • @r.k.vignesh7832
    @r.k.vignesh7832 2 місяці тому

    8:24 do you have any sources for this? Are you estimating from the FineWeb 15T dataset, but fully deduplicated, to achieve the 3T tokens across the Internet number?

  • @borb5353
    @borb5353 3 місяці тому +1

    isnt this kinda why RAG is so effective ?
    i mean if u just let the LLM add new entries to the database, sort through the database, go over what makes sense and what doesnt, and ofc use the entries in the database to give it more relevant tokens then it would just keep improving as the quality of the tokens in the input improves naturally over time through the process

  • @Jimpipecigarbear
    @Jimpipecigarbear 2 місяці тому

    It is interesting pushing AI to make two grammer check systems happy. Through standing my ground in implementing. I am dedicated to avoiding the best version suggestion of grammerly. Especially since best version is subjective and through focus of making it an assistant instead of an writer.

  • @chasisaac
    @chasisaac 3 місяці тому

    So I have been mostly using Claude for writing. Why I have found is have it rewrite it three times before I even look at it and it is so much better.
    My primary prompt has gone from two or 300 words to 4000. I’m wondering if I’m just tossing words at it.

    • @carultch
      @carultch 3 місяці тому

      You mean word gambling, not writing.

    • @chasisaac
      @chasisaac 3 місяці тому

      @@carultch Not really I do a step by step chain of thought reasoning. So it goes does take then looks for help. Then move onto next task. I use Claude projects. Add in info and move re run. I am getting better results. Admittedly it might be better with an agent.

  • @DiamondDepthYT
    @DiamondDepthYT 2 місяці тому

    Prompt engineering is actually kinda like software engineering.. kinda. This is just from my current knowledge, but:
    At least for me- I come up with an algorithm to give to the AI, and then I sorta debug what it spits out.

  • @SpentAmbitionDrain
    @SpentAmbitionDrain 3 місяці тому +17

    This is such a pointless job. People who choose to do this will hate themselves really soon. When I hit a wall as a dev I can get myself out of the bind in any number of deterministic ways. When you can't convince an LLM to spit out token in the right order you will wish you became an electrician really soon.

    • @denjamin2633
      @denjamin2633 3 місяці тому +9

      Bra for 300k a year I can put up with a bad job for a long ass time.

    • @hmmm....1910
      @hmmm....1910 3 місяці тому

      Now imagine doing the same job for 9k per year!​@@denjamin2633

    • @notsojharedtroll23
      @notsojharedtroll23 3 місяці тому +1

      ​@@denjamin2633 yep, imagone doing something as simple as chstting to a basicallu groundbreking tech to the world and getting paid a 300k USDper year salary 🤣🤣🤣

  • @AD-zj7ck
    @AD-zj7ck 3 місяці тому

    "The entire usable internet is 3 trillion tokens" from where did that number came from?

    • @bycloudAI
      @bycloudAI  3 місяці тому +3

      llama3 herd of models paper

  • @Avegeania
    @Avegeania 3 місяці тому

    I just use gemini to prepare prompts about what I need and feed those to gpt or Claude.

  • @kombuchamooch
    @kombuchamooch 3 місяці тому

    The Rs in the word strawberry one is funny because just about everyone posting that didn't even pluralize Rs correctly, that's probably not helping.

  • @VacouslyTrivial
    @VacouslyTrivial 3 місяці тому +4

    4:03(i'm not actually an LLM😂😂)

    • @KCstark.
      @KCstark. 3 місяці тому

      Ai denying 😢

  • @CHURCHISAWESUM
    @CHURCHISAWESUM 3 місяці тому

    Thanks for the AI slop bro

  • @kyosokutai
    @kyosokutai 2 місяці тому

    ... I always just called it AI-Wrangling. Because it's next to impossible to get the results you want and it'd be faster to just learn to draw/code/write.

  • @Edgard28Fe
    @Edgard28Fe 3 місяці тому

    uff, my brain got depleted of brain juice just trying to follow and grasp all that info 🤯

  • @charmec8267
    @charmec8267 3 місяці тому +1

    Was this video narrated by an AI voice? It tends to end a sentence on a very flippant intonation that's quite offputting.

  • @BHBalast
    @BHBalast 3 місяці тому

    My insight from all of this is just that generating more tokens about reasoning and revelant topic helps to move the "probable distribution" closer to a valid answer so thats the technical way all of this methods work. Some of the reasoning of humans also works that way or it just looks that way, I'd say it's more reasoning from analogy than from first principles, I wonder if LLM's just don't know how to reason from first principles and when to reason from analogy. Humans also struggle with that. 😅

  • @joecavanagh1297
    @joecavanagh1297 3 місяці тому

    That thumbnail lol.

  • @TrisAF
    @TrisAF 3 місяці тому

    As the technology advances prompt engineering will probably go away right? I mean you'll need to be able to communicate as well as you do to your employees or your coworkers to it but that's the point of all the improvement to AI over the years right? Every milestone is an improvement and how well it recognizes natural language with functionality to improv gathering context and intent being cornerstones of any project that utilizes AI. To put it simply, the functionalities we see in the movie Her dont seem that far fetched as it becomes more and more realistic. I'm not arguing it ever will truly be conscious like the movie, but its ability to be spoken to naturally like a human would just seems to make prompt engineering obsolete.

  • @codycast
    @codycast 3 місяці тому +4

    I still don’t get “prompt engineering”. Maybe I’m just a natural? But I ask ChatGPT to do something and it does it. I don’t need to hire an engineer to specially formulate my question.

    • @akirachisaka9997
      @akirachisaka9997 3 місяці тому +4

      From my understanding, prompt engineering is more about “how do we package the user’s question so that when we feed it into an LLM, the LLM can respond in a way that matches what the user want, even when the user have no idea what the fuck is an LLM”.
      So yeah, when using ChatGPT, what you typed into the chat box isn’t what’s directly being fed into the model. There still needs to be a wrapper around it, something like “you are a helpful AI assistant blah blah blah”. So in this way, the prompt engineering is already being mostly done.
      The name “prompt engineering” likely comes from “specification engineering”. Which is about “how do we craft a ‘prompt’ that we can feed into a team of human coders so that the systems they make are actually useful?”
      Yeah calling it “engineering” is a bit weird. But I guess it’s because how it’s mostly writing engineering documentation, which is like… 90% of what engineers do anyway?

    • @codycast
      @codycast 3 місяці тому

      @@akirachisaka9997 I’d love to get more info as to what happens to my prompt behind the scenes before the LLM gets it
      Like with google we know they had all this racist nonsense going on to make black female popes. So it would be nice if the built in wrappers were more open.

    • @NighttimeJuneau
      @NighttimeJuneau 3 місяці тому +3

      I guess because LLMs are trained on human data, if you’re already good at asking questions and explaining problems to humans, you’ll also be good at prompt engineering.

    • @reinerheiner1148
      @reinerheiner1148 3 місяці тому +1

      Well, the better the prompt, the better the answer. Thats because an LLM cannot read your mind, therefore only has the context from your prompt (and the chat history if there is any). Precise language also is key for the LLM to know what precisely you want. But for easy questions, thats not needed.
      Prompt engineering is often required to make sure that a specific output is (hopefully lol) reproducable, so that a specific prompt at a specific part in a workflow can do one specific thing reliably. Another thing is making system prompts, that make the LLM as good as they can get, for answering the prompts the subsequent user gives the LLM. I have seen widely different output from an LLM based on how I prompted it to do the same thing, its like night and day. Of course, its pointless for any prompt that an LLM can handle easily

    • @xponen
      @xponen 3 місяці тому

      if we describe X using common parlance then it gave answers that is different than if we describe X using term specific to X, especially in science & math. Basically, we have to present the correct term to extract the correct answer to X. We don't know that term, so we have to feed-back any word that came out of LLM in hope to trigger context that is closer to X.

  • @ai-programming-p4v
    @ai-programming-p4v 3 місяці тому

    Can you reduce motion in a video? I was watching you cause you know stuff and explain stuff and not for wobbly animation

  • @karthikeyank2587
    @karthikeyank2587 3 місяці тому +4

    Some companies work as wrapper of openai with prompts of themselves,it is really difficult to find whether it is a chatbot is gpt wrapper or a real proprietor model created on own.
    Ri8 now not everyone is trying prompt injection to check whether to check whether it is wrapper or own/fine-tuned model.
    With the use of Guardrails it is easy for wrapper based company
    Tldr: is there a way to find a chatbot a company promise it is a prosperity LLM developed on own is authentic or it is just a gpt wrapper?

  • @mariokotlar303
    @mariokotlar303 3 місяці тому

    Are you saying that gpt4 has been o1 intelligent all along, we just didn't know how to squeeze it out of it?

    • @ZazeLove
      @ZazeLove 3 місяці тому

      gpt 4 works better in many cases...

    • @emperorpalpatine6080
      @emperorpalpatine6080 3 місяці тому

      You know that o1 is just fancy prompting behind , on gpt 4 , right ?😂

    • @ZazeLove
      @ZazeLove 3 місяці тому

      @@emperorpalpatine6080 That's not correct, it uses a completely different approach. it still has aspects of the older systems, but that will change as more good data is accumulated.
      There are some instances where the chain of thought and review process in o-1 have advantages, Specifically its ability to understand and reason through why issues are happening is rather handy, but some times it enters a recursive loop of failed logic, much like the 4-o model.
      There's advantages and dis advantages to every approach, and some use cases are not something an AI can handle. ( yet )
      I am aware of a lot more then the average person who doesn't spend any time researching or using these models, I code with them for hours, some days I'm sitting around for 10 plus hours working on projects with GPT running script generation on the side virtually the entire time, I have to Test the code to make sure it compiles correctly, But It depends on the person, and the use case, I train the god damn model quite often.

  • @naninano8813
    @naninano8813 3 місяці тому +1

    always tip your llm.

  • @zalzalahbuttsaab
    @zalzalahbuttsaab 3 місяці тому

    o1 can't sort three characters into alphabetical order. there i said it.

    • @charlesabju907
      @charlesabju907 2 місяці тому

      People have said this: its an architect not an intern.
      It kinda trolls when prompted for excedingly simple questions. I use o1 daily only for super complex questions and its responses are much better than 4o's

  • @isbestlizard
    @isbestlizard 3 місяці тому

    Read it again! I just like to hear the story!

  • @Outspokenknight
    @Outspokenknight 3 місяці тому +1

    Ask how it many Ps are in PooPooPeePee

    • @xponen
      @xponen 3 місяці тому +2

      Q: how many Ps are in PooPooPeePee
      A: There are four "P" letters in "PooPooPeePee."
      Q: Please read the question again
      A: I see what you mean! There are six "P" letters in "PooPooPeePee." Thanks for your patience!

  • @Dmartin23941
    @Dmartin23941 3 місяці тому

    r/singularity bros be like

  • @betterlit
    @betterlit 3 місяці тому

    what's with your shadow bouncing

  • @abhishekak9619
    @abhishekak9619 3 місяці тому

    they are doing it. they are training it on generating the next n tokens. i have been saying this for a long time.

  • @MrSomethingred
    @MrSomethingred 3 місяці тому

    Wait, dont tell me people are paying from prompt templates? thats embarrassing

  • @karthikeyank2587
    @karthikeyank2587 3 місяці тому +3

    Nowadays am really feed up following the trend in this LLM era
    Prompt engineering is been there long ago am speaking about things,every day some model is fine tuned and released in hugging face
    Openai o1 came and Qwen 2.5 came now llama 3.2 in next 15-30 Google or Microsoft or nvidia will release a new model this cycle goes on and on
    Btw ad from 1:49 to 3:19

  • @ij9375
    @ij9375 3 місяці тому

    Best AI channel ❤

  • @PixelSubstream
    @PixelSubstream 2 місяці тому

    You think it would be more valuable to make the AI behave in the intended way instead of hoping someone is going to figure out how to ask something exactly right every time

  • @Cheesystreams2
    @Cheesystreams2 3 місяці тому +3

    why go from black background to white and back and forth, invert the damn research papers!

  • @farpurple
    @farpurple 2 місяці тому

    Use pyrhon + nltk or something..

  • @serenditymuse
    @serenditymuse 3 місяці тому +3

    That HubSpot thing that is widely pushed by AI content creators is pretty much useless.

  • @TheAkdzyn
    @TheAkdzyn 3 місяці тому

    This is good for non English speaking countries

  • @Joseph-kd9tx
    @Joseph-kd9tx 3 місяці тому

    5:44 real

  • @NotAGeySer
    @NotAGeySer 3 місяці тому

    Huh!

  • @astrovation3281
    @astrovation3281 2 місяці тому

    it's not a job the average person should learn how to be a prompt engineer

  • @cleverhype
    @cleverhype 3 місяці тому +1

    Mega ad

  • @anywallsocket
    @anywallsocket 3 місяці тому

    The nature of the answer has everything to do with the nature of the question. It’s the faulty assumption that what you’ve asked, is what you’ve asked, that leads us to believe better prompts yield better answers.

  • @nikity_
    @nikity_ 3 місяці тому

    😍

  • @casey206969
    @casey206969 3 місяці тому +1

    stop uptalking

  • @SHIROmori3
    @SHIROmori3 2 місяці тому

    Dude! talk slower, leave slides longer on screen. Busta Rhymes here....

  • @ginamejia-b4b
    @ginamejia-b4b 3 місяці тому +1

    Your high pitch singing is annoying

  • @Zuluknob
    @Zuluknob 3 місяці тому

    "prompting shortly is always a bad idea" not when it comes to image generation. ... flux ... "pretty, ugly. tame, wild." generate. generate. generate. generate. generate. generate. generate. generate. generate. generate. generate. generate.