Multimodal RAG with GPT-4-Vision and LangChain | Retrieval with Images, Tables and Text

Поділитися
Вставка
  • Опубліковано 9 чер 2024
  • In this video I will show you how to do retrieval with a multivector retriever. First we will disucss how we can do retrieval with images, tables and text by creating summaries for each element.
    Then we will go through the jupyter notebook.
    Code: github.com/Coding-Crashkurse/...
    Timestamps:
    0:00 Theory
    1:27 Code walkthrough

КОМЕНТАРІ • 29

  • @nmstoker
    @nmstoker 8 днів тому

    Nice video. Would be nice to go into a few more examples/use cases to more strongly illustrate why multimodal RAG is useful

  • @micbab-vg2mu
    @micbab-vg2mu 6 місяців тому

    thank you

  • @arishasaeed
    @arishasaeed 7 годин тому

    How can I use pinecone instead of chroma here?

  • @Slimshady68356
    @Slimshady68356 6 місяців тому

    Nicely explained ,Subscribed🎉

  • @amortalbeing
    @amortalbeing 6 місяців тому

    thanks❤

  • @saurabhjain507
    @saurabhjain507 6 місяців тому

    Will this part of partitioning work on Azure? How do you read pdf from a storage container?

    • @codingcrashcourses8533
      @codingcrashcourses8533  6 місяців тому

      I have not tried this yet. I would use the Azure SDK, but not sure if that works the same as reading the file from the local filesystem

    • @saurabhjain507
      @saurabhjain507 6 місяців тому

      @@codingcrashcourses8533 PDFs stored as blobs on Azure are different than reading locally. I have tried using langchain but was not able to read it. I then used pypdf to read the pdf as a streaming object.

  • @chakerayachi8468
    @chakerayachi8468 Місяць тому

    nicely explained and nice informations as always but i have a question my files are stored in azure blob storage i am getting tghem throw blob loader does implementing the multimodal works with them?

    • @codingcrashcourses8533
      @codingcrashcourses8533  Місяць тому +1

      I don´t know to be honest, but I think it should be possible. If not maybe try to get the files directly with the Azure SDK

    • @chakerayachi8468
      @chakerayachi8468 Місяць тому

      @@codingcrashcourses8533 as always thanks for replying to my comments my mentor

  • @AdarshMamidpelliwar
    @AdarshMamidpelliwar 3 місяці тому

    Can we show the images as response along with relevant text as response based on the prompt passed.

    • @codingcrashcourses8533
      @codingcrashcourses8533  3 місяці тому

      Yes, but I would probably do that different. Maybe with a different embedding model. But to be honest, I can not good idea out of the box.

  • @akmr0079
    @akmr0079 3 місяці тому

    cant we get image also if we use vision model in the chain?

  • @yazanrisheh5127
    @yazanrisheh5127 6 місяців тому

    why did you use chiain.invoke and not .run or apply or batch? Sometimes in your videos you use run and sometimes invoke. How od you know when to use when and whats the difference?

    • @codingcrashcourses8533
      @codingcrashcourses8533  6 місяців тому +1

      I thought about using batch and think its probably better, but I tried to keep it simple and just use a loop for every call.
      The difference between run and invoke is the chain. I try to use Language expression language only in my newer videos and invoke is the implementation of the runnable interface, while run is the implementation of the (deprecated) chain interface

  • @muzammilnizamani8585
    @muzammilnizamani8585 6 місяців тому

    will this output images with text as well?

  • @egitaufiqnoor3612
    @egitaufiqnoor3612 Місяць тому

    how to store vectore created to local? so ican used again later

    • @codingcrashcourses8533
      @codingcrashcourses8533  Місяць тому

      Faiss and chroma offer methods to do that. You will find that in the langchain docs

  • @ajeetojha3745
    @ajeetojha3745 2 місяці тому

    Can you pls share the notebook

  • @Slimshady68356
    @Slimshady68356 6 місяців тому

    Hi Markus ,I am having problem with downloading tesseract ,the download is really slow , do you have any link to tesseract

    • @codingcrashcourses8533
      @codingcrashcourses8533  6 місяців тому +1

      digi.bib.uni-mannheim.de/tesseract/ Hello Zaid, this is another Link I used before. Hope that helps! Best regards

    • @Slimshady68356
      @Slimshady68356 6 місяців тому

      @@codingcrashcourses8533 Thanks Markus!!
      😊