What is Retrieval-Augmented Generation (RAG)?

Поділитися
Вставка
  • Опубліковано 21 тра 2024
  • Try RAG with watsonx → ibm.biz/BdMsRT
    Learn more about RAG→ ibm.biz/BdMsRt
    Large language models usually give great answers, but because they're limited to the training data used to create the model. Over time they can become incomplete--or worse, generate answers that are just plain wrong. One way of improving the LLM results is called "retrieval-augmented generation" or RAG. In this video, IBM Senior Research Scientist Marina Danilevsky explains the LLM/RAG framework and how this combination delivers two big advantages, namely: the model gets the most up-to-date and trustworthy facts, and you can see where the model got its info, lending more credibility to what it generates.
    Get started for free on IBM Cloud → ibm.biz/sign-up-now
    Subscribe to see more videos like this in the future → ibm.biz/subscribe-now

КОМЕНТАРІ • 375

  • @xzskywalkersun515
    @xzskywalkersun515 6 місяців тому +490

    This lecturer should be given credit for such an amazing explanation.

    • @cosmicscattering5499
      @cosmicscattering5499 3 місяці тому +3

      I was thinking the same, she explained this so clearly.

    • @tariqmking
      @tariqmking 2 місяці тому +1

      Yes this was excellently explained, kudos to her.

    • @brianmi40
      @brianmi40 Місяць тому +6

      Or at least credit for being able to write backwards!

    • @victoriamilhoan512
      @victoriamilhoan512 14 днів тому

      The connection between a human answering a question in real life vs how LLMs (with or without RAG) do it was so helpful!

  • @vt1454
    @vt1454 6 місяців тому +316

    IBM should start a learning platform. Their videos are so good.

    • @XEQUTE
      @XEQUTE 5 місяців тому +6

      i think they already do

    • @srinivasreddyt9555
      @srinivasreddyt9555 2 місяці тому

      Yes, they have it already. UA-cam.

    • @siddheshpgaikwad
      @siddheshpgaikwad Місяць тому +2

      Its mirrored video, she wrote naturally and video was mirrored later

    • @Hossam_Ahmed_
      @Hossam_Ahmed_ Місяць тому

      They have skill build but not videos at least most of the content

    • @CaptPicard81
      @CaptPicard81 Місяць тому

      They do, I recently attended a week long AI workshop based on an IBM curriculum

  • @ghtgillen
    @ghtgillen 7 місяців тому +54

    Your ability to write backwards on the glass is amazing! ;-)

    • @jsonbourne8122
      @jsonbourne8122 6 місяців тому +21

      They flip the video

    • @Paul-rs4gd
      @Paul-rs4gd 4 місяці тому +9

      @@jsonbourne8122 So obvious, but I did not think of it. My idea was way more complicated!

    • @aykoch
      @aykoch 4 дні тому

      They're almost always left-handed as well...

  • @jordonkash
    @jordonkash 3 місяці тому +29

    4:15 Marina combines the colors of the word prompt to emphasis her point. Nice touch

  • @natoreus
    @natoreus 12 днів тому +4

    I'm sure it was already said, but this video is the most thorough, simple way I've seen RAG explained on YT hands down. Well done.

  • @geopopos
    @geopopos 2 місяці тому +42

    I love seeing a large company like IBM invest in educating the public with free content! You all rock!

  • @vikramn2190
    @vikramn2190 8 місяців тому +30

    I believe the video is slightly inaccurate. As one of the commenters mentioned, the LLM is frozen and the act of interfacing with external sources and vector datastores is not carried out by the LLM.
    The following is the actual flow:
    Step 1: User makes a prompt
    Step 2: Prompt is converted to a vector embedding
    Step 3: Nearby documents in vector space are selected
    Step 4: Prompt is sent along with selected documents as context
    Step 5: LLM responds with given context
    Please correct me if I'm wrong.

    • @DJ-lo8qj
      @DJ-lo8qj Місяць тому

      I’m not sure. Looking at OpenAI documentation on RAG, they have a similar flow as demonstrated in this video. I think the retrieval of external data is considered to be part of the LLM (at least per OpenAI)

    • @PlaytimeEntertainment
      @PlaytimeEntertainment Місяць тому +1

      I do not think retrieval is part of LLM. LLM is the best model at the end of convergence after training. It can't be modified rather after LLM response you can always use that info for next flow of retrieval

  • @ericadar
    @ericadar 5 місяців тому +51

    Marina is a talented teacher. This was brief, clear and enjoyable.

  • @trugoku
    @trugoku 3 дні тому

    I love these types of conversations.
    Great video, dialog and explanation breakdown.

  • @maruthuk
    @maruthuk 7 місяців тому +20

    Loved the simple example to describe how RAG can be used to augment the responses of LLM models.

  • @ntoscano01
    @ntoscano01 4 місяці тому +21

    Very well explained!!! Thank you for your explanation of this. I’m so tired of 45 minute UA-cam videos with a college educated professional trying to explain ML topics. If you can’t explain a topic in your own language in 10 minutes or less than you have failed to either understand it yourself or communicate effectively.

  • @javi_park
    @javi_park 3 місяці тому +31

    hold up - the fact that the board is flipped is the most underrated modern education marvel nobody's talking about

    • @RiaKeenan
      @RiaKeenan 3 місяці тому

      I know, right?!

    • @euseikodak
      @euseikodak 3 місяці тому +4

      Probably they filmed it in front of a glass board and flipped the video on edition later on

    • @politicallyincorrect1705
      @politicallyincorrect1705 3 місяці тому

      Filmed in front of a non-reflective mirror.

    • @TheTomtz
      @TheTomtz Місяць тому

      Just simply write on a glass board ,record it from the other side and laterally flip the image! Simple aa that.. and pls dont distract people from the contents being lectured by thinkin about the process behind the rec🤣

    • @thewallstreetjournal5675
      @thewallstreetjournal5675 Місяць тому

      Is the board fliped or has she been flipped?

  • @TheAllnun21
    @TheAllnun21 5 місяців тому +16

    Wow, this is the best beginner's introduction I've seen on RAG!

  • @aam50
    @aam50 5 місяців тому +17

    That's a really great explanation of RAG in terms most people will understand. I was also sufficiently fascinated by how the writing on glass was done to go hunt down the answer from other comments!

  • @ReflectionOcean
    @ReflectionOcean 5 місяців тому +21

    1. Understanding the challenges with LLMs - 0:36
    2. Introducing Retrieval-Augmented Generation (RAG) to solve LLM issues - 0:18
    3. Using RAG to provide accurate, up-to-date information - 1:26
    4. Demonstrating how RAG uses a content store to improve responses - 3:02
    5. Explaining the three-part prompt in the RAG framework - 4:13
    6. Addressing how RAG keeps LLMs current without retraining - 4:38
    7. Highlighting the use of primary sources to prevent data hallucination - 5:02
    8. Discussing the importance of improving both the retriever and the generative model - 6:01

  • @m.kaschi2741
    @m.kaschi2741 5 місяців тому +5

    Wow, I opened youtube coming from the ibm blog just to leave a comment. Clearly explained, very good example, and well presented as well!! :) Thank you

  • @Lucildor
    @Lucildor 3 місяці тому

    Please keep all these videos coming! They are so easy to understand and straightforward. Muchas gracias!

  • @GregSolon
    @GregSolon 3 місяці тому

    One of the easiest to understand RAG explanations I've seen - thanks.

  • @jyhherng
    @jyhherng 6 місяців тому +5

    this let's me understand why the embeddings used to generate the vectorstore is a different set from the embeddings of the LLM... Thanks, Marina!

  • @444Yielding
    @444Yielding Місяць тому +3

    This video is highly underviewed for as informative as it is!

  • @kingvanessa946
    @kingvanessa946 3 місяці тому +1

    For me, this is the most easy-to-understand video to explain RAG!

  • @projectfocrin
    @projectfocrin 6 місяців тому +5

    Great explanation. Even the pros in the field I have never seen explain like this.

  • @redwinsh258
    @redwinsh258 6 місяців тому +22

    The interesting part is not retrieval from the internet, but retrieval from long term memory, and with a stated objective that builds on such long term memory, and continually gives it "maintenance" so it's efficient and effective to answer. LLMs are awesome because even though there are many challenges ahead, they sort of give us a hint of what's possible, without them it would be hard to have the motivation to follow the road

  • @paulaenchina
    @paulaenchina 4 місяці тому +1

    This is the best explanation I have seen so far for RAG! Amazing content!

  • @hamidapremani6151
    @hamidapremani6151 2 місяці тому +1

    The explanation was spot on!
    IBM is the go to platform to learn about new technology with their high quality content explained and illustrated with so much simplicity.

  • @past_life_project
    @past_life_project 4 місяці тому

    I have watched many IBM videos and this is the undoubtedly the best ! I will be searching for your videos now Marina!

  • @janhorak8799
    @janhorak8799 2 місяці тому +10

    Did all the speakers have to learn how to write in a mirrored way or is this effect reached by some digital trick?

    • @VlogBySKSK
      @VlogBySKSK Місяць тому

      There is a digital mirroring technique which is used to show the content this way...

    • @mao-tse-tung
      @mao-tse-tung 29 днів тому +3

      She was right handed before the mirror effect

  • @rujmah
    @rujmah 2 місяці тому

    Brilliant explanation and illustration. Thanks for your hard work putting this presentation together.

  • @Anubis2828
    @Anubis2828 2 місяці тому

    Great, simple, quick explanation

  • @rsu82
    @rsu82 9 днів тому

    good explanation, it's very easy to understand. this video is the first one when I search RAG on UA-cam. great job ;)

  • @evaiintelligence
    @evaiintelligence Місяць тому

    Marina has done a great job explaining LLM and RAGs in simple terms.

  • @francischacko3627
    @francischacko3627 29 днів тому

    perfect explanation understood every bit , no lags kept it very interesting ,amazing job

  • @vnaykmar7
    @vnaykmar7 5 місяців тому +2

    Such an amazing explanation. Thank you ma'am!

  • @user-cd6hp5kc1n
    @user-cd6hp5kc1n 7 місяців тому +16

    The ability to write backwards, much less cursive writing backwards, is very impressive!

    • @IBMTechnology
      @IBMTechnology  7 місяців тому +7

      See ibm.biz/write-backwards

    • @jsonbourne8122
      @jsonbourne8122 6 місяців тому

      Left hand too!

    • @NishanSaliya
      @NishanSaliya 6 місяців тому

      @@IBMTechnology Thanks .... I was reading comments to check for an answer for that question!

  • @Aryankingz
    @Aryankingz 7 місяців тому +3

    That's what Knowledge graphs are for, to keep LLMs grounded with a reliable source and up-to-date.

  • @sawyerburnett8319
    @sawyerburnett8319 4 місяці тому +1

    Wow, having a lightbulb moment finally after hearing this mentioned so often. Makes more sense now!

  • @rockochamp
    @rockochamp 5 місяців тому +1

    very well executed presentation.
    i had to think twice about how you can write in reverse but then i RAGed my system 2 :)

  • @mstarlingc
    @mstarlingc 5 місяців тому +1

    Pretty simple explanation, thank you

  • @Shailendrashail
    @Shailendrashail 8 місяців тому +1

    Good Explanation of RAG. Thanks for sharing.

  • @rvssrkrishna2
    @rvssrkrishna2 2 місяці тому

    Very precise and exact information on RAG in a nutshell. Thank you for saving my time.

  • @rafa1rafa
    @rafa1rafa 5 місяців тому +2

    Great explanation! The video was very didactic, congratulations!

  • @kallamamran
    @kallamamran 4 місяці тому +2

    We also need the models to cross check their own answers with the sources of information before printing out the answer to the user. There is no self control today. Models just say things. "I don't know" is actually a perfectly fine answer sometimes!

  • @TheMsksk
    @TheMsksk 9 місяців тому

    Great video as always. Thanks for sharing.

  • @PaulGrew-wl7mh
    @PaulGrew-wl7mh Місяць тому

    An amazing explanation that made RAG understandable in about 4:23 minutes!

  • @xdevs23
    @xdevs23 2 місяці тому +5

    The entire video I've been wondering how they made the transparent whiteboard

  • @bdouglas
    @bdouglas Місяць тому

    That was excellent, simple, and elegant! Thank you!

  • @HimalayJoriwal
    @HimalayJoriwal 2 місяці тому

    Best explanation so far from all the content on internet.

  • @kunalsoni7681
    @kunalsoni7681 6 місяців тому

    Thanks for letting us know about this feature of LLM :)

  • @toenytv7946
    @toenytv7946 2 місяці тому +1

    Great down the rabbit hole video. Very deep and understandable. IBM academy worthy in my opinion.

  • @JasonVonHolmes
    @JasonVonHolmes 2 місяці тому

    This was explained fantastically.

  • @mohamadhijazi3895
    @mohamadhijazi3895 Місяць тому

    The video is short and consice yet the delivery is very elegant. She might be the best instructor that have teached me. Any idea how the video was created?

  • @gaemrpaterso-ri2jd
    @gaemrpaterso-ri2jd 9 місяців тому

    Great video, you guys should do one on promising tech industries

  • @khalidelgazzar
    @khalidelgazzar 5 місяців тому +2

    Great explanation. Thank you!😊

  • @laurentpastorelli1354
    @laurentpastorelli1354 4 місяці тому +1

    Super good and clear, well done!

  • @shashankshekharsingh9336
    @shashankshekharsingh9336 22 дні тому

    very good and clear explanation

  • @johnmccullough7084
    @johnmccullough7084 6 місяців тому +1

    Appreciate the succinct explanation. 👍

  • @user-im6ub3sf6m
    @user-im6ub3sf6m 3 місяці тому

    Great explanation with an example. Thank you

  • @user-hk5dk9rb6p
    @user-hk5dk9rb6p 4 місяці тому +1

    Fantastic video and explanation. Thank you!

  • @Kekko400D
    @Kekko400D 3 місяці тому

    Fantastic explanation, proud to be an IBMer

  • @zuzukouzina-original
    @zuzukouzina-original 4 місяці тому

    Very clear explanation, much respect 🫡

  • @thomasbrowne6649
    @thomasbrowne6649 Місяць тому

    This is excellent and I hope IBM does well in this space. We need a reliable, non-hype vendor.

  • @prasannakulkarni5664
    @prasannakulkarni5664 Місяць тому

    the color coding on your whiteboard is really apt here !

  • @AdarshKumar-kx2cn
    @AdarshKumar-kx2cn 3 місяці тому

    Beautifully explained....thanks

  • @eddisonlewis8099
    @eddisonlewis8099 4 місяці тому

    AWESOME EXPLANATION OF THE CONCEPT RAG

  • @lauther_27
    @lauther_27 5 місяців тому +1

    Amazing video, thanks IBM ❤

  • @preciousrose2715
    @preciousrose2715 Місяць тому

    This was such an amazing explanation!

  • @421sap
    @421sap 6 місяців тому

    Thank you, Marina Danilevsky ....

  • @alexiojunior7867
    @alexiojunior7867 Місяць тому

    wow this was an amazing Explanation ,very easy to understand

  • @ashwinkumar675
    @ashwinkumar675 29 днів тому

    This is so well explained! Thank you 👍🏻✅

  • @afshinkarimi2382
    @afshinkarimi2382 8 місяців тому

    Great video. Thanks for sharing

  • @star2k279
    @star2k279 4 місяці тому +1

    Thank you for such a great explanation.

  • @mrhassell
    @mrhassell Місяць тому

    RAG combines the generative power of LLMs with the precision of specialized data search mechanisms, resulting in nuanced and contextually relevant responses.

  • @neutron417
    @neutron417 5 місяців тому +2

    From which corpus/database are the documents retrieved from? Are they up-to date? and how does it know the best documents to select from a given set?

  • @user-uk9mt4ue6w
    @user-uk9mt4ue6w 5 місяців тому +1

    Все толково, четко и понятно. Респект автору.

  • @oieieio741
    @oieieio741 5 місяців тому

    Very Helpful! Great explanation. thx IBM

  • @MraM23
    @MraM23 3 місяці тому

    Great lessons! Nice of you to step out 🙃 and make such engaging and educative content This is a very useful in helping us in critical thinking. Thank you for sharing this video. 👍
    Current ai models may impose neurotypical norms and expectations based on current data trained on . 🤔
    Curious to see more on how IBM approach the challenges and limitations of Ai

  • @user-bo1kv5zy3w
    @user-bo1kv5zy3w 7 місяців тому

    Awesome explanation. Love you.

  • @AntenorTeixeira
    @AntenorTeixeira 5 місяців тому

    That's the best video about RAG that I've watched

  • @aniket_mishr
    @aniket_mishr Місяць тому

    The explanation was very good 💯.

  • @yashkhorania3726
    @yashkhorania3726 21 день тому

    very nicely explained

  • @mayankbumb7272
    @mayankbumb7272 12 днів тому

    Great explanation

  • @ericmcnally5128
    @ericmcnally5128 2 місяці тому

    This is a fantastic lesson video.

  • @AC-xd7sw
    @AC-xd7sw 4 місяці тому

    Insightful, please more video like this

  • @randomforest_dev
    @randomforest_dev Місяць тому

    Very good explanation!

  • @kevinmulligan2006
    @kevinmulligan2006 2 місяці тому

    tokens as a [word] is what I'm working on right now (solo, self learning LLM techniques), this video helped me realize how the model doesn't know what it's outputting obviously, but AI-AI is different, so building tokens that have dimensional vectors that process in a separate model, can be used for explainable AI.

    • @kevinmulligan2006
      @kevinmulligan2006 2 місяці тому

      meaning a separate model processes the response itself, meta, it's for building evolution learning. AI-AI machine learning, you need a way to configure in between the iterations.

  • @deltawhiplash1614
    @deltawhiplash1614 18 днів тому

    This is a really good video thank you for sharing this knowledge

  • @stanislavzayarsky
    @stanislavzayarsky 3 місяці тому

    Finally, we got a clear explanation!

  • @siddharth4251
    @siddharth4251 Місяць тому

    Amazing explanation, finally i understand it.

  • @DaveEtchells
    @DaveEtchells 5 місяців тому +1

    *Excellent* explanation, you gave me all the key concepts in one shot.
    I gather that the retrieval could be in various forms, for instance a vector database in addition to direct text from internet searches?

    • @scottt9382
      @scottt9382 5 місяців тому

      a RAG target can be any associated data store (PDF repo, Sharepoint, Google Drive,..) that can be accessed via a query - the LLM has a semantic understanding of the prompt and the queries are the output of the LLM

  • @lewi594
    @lewi594 4 місяці тому +1

    This is brilliant and concise, helped make sense of a complex subject..
    Can this be implemented in a small environment with limited computing? Such that the retriever only has access to a closed data source

  • @sumedhaj9017
    @sumedhaj9017 2 місяці тому

    Amazing explanation! Thank you:)

  • @BooleanDisorder
    @BooleanDisorder 5 місяців тому +1

    Thank you for these videos. Makes it much easier to nagivate this new AI-ra of machine learning.

  • @sudhakarveeraraghavan5832
    @sudhakarveeraraghavan5832 Місяць тому

    Very well explained and it is easily understandable to non AI person as well. Thanks.

  • @Junglytics
    @Junglytics 3 місяці тому

    Great video, excellent explanation!

  • @terencelewis4985
    @terencelewis4985 3 місяці тому

    Excellent explanation!

  • @rahulberry4806
    @rahulberry4806 26 днів тому

    thanks for the great explanation

  • @aneesarom
    @aneesarom 6 днів тому

    Best explanation ever

  • @gbluemink
    @gbluemink 3 місяці тому +1

    So the question I have here is when I have an answer from my LLM but not the Rag data, what is the response to the user? "I don't know" or the LLM response that may be out of date or without a reliable source? Looks like a question for an LLM :)

  • @katsunoi
    @katsunoi 5 місяців тому

    nice video - great explanation!

  • @sprintwithcarlos
    @sprintwithcarlos 6 місяців тому

    Great explanation!