Meet KAG: Supercharging RAG Systems with Advanced Reasoning

Поділитися
Вставка
  • Опубліковано 10 лют 2025

КОМЕНТАРІ • 66

  • @engineerprompt
    @engineerprompt  Місяць тому +7

    If you are interested in learning more about Advanced RAG systems, checkout my RAG Beyond Basics Course: prompt-s-site.thinkific.com/courses/rag

  • @donb5521
    @donb5521 Місяць тому +11

    The premise of Knowledge Augmented Generation is promising, but the current KAG code bases failed to deliver today. The TLDR version is that ultimately I saw no notes or edges created in Neo4j. Even weirder is that in spite of there being no graph it was still giving me results in the UI. (the UI is not open source and appears to be locked down)
    Ultimately the config needs to become more solid and consistent -- and there needs to be agreement between the OpenSPG/openspg and OpenSPG/KAG development teams on whether Ollama is supported.
    An odd mix of Java and Python. Hopefully this gets straightened up soon.
    Prompt Engineering... Normally love your stuff. What would be helpful as a starting point is a Jupyter Notebook from OpenSPG that walks through (and validates) the pipeline step by step. A follow-up would be a reproducible and well documented evaluation against other solutions - LazyGraphRAG, LightRAG / nanorag, etc.

  • @SullyOrchestration
    @SullyOrchestration Місяць тому +16

    I've tried this and unfortunately it's way too slow (30 seconds for a query with a 2 sentence output!) and does not produce its 'logical form solver' transparently enough. We don't know what chunks were retrieved or from where. Unfortunately it's still quite a way from being usable in a practical app.

    • @KoteikiJan
      @KoteikiJan Місяць тому +2

      It depends on the use case! Sometimes it's acceptable to wait a few minutes, or maybe even hours.

    • @MatichekYoutube
      @MatichekYoutube Місяць тому

      @@KoteikiJan can you write an example in which case would be for example ok to wait for hours? maybe for lawyers?

    • @KoteikiJan
      @KoteikiJan Місяць тому +1

      @ yes, court case analysis is one, and any other kinds of analysis that can be executed overnight. Also generating book text, or various kinds of documentation. Of course, being able to iterate fast is an advantage, but sacrificing speed for quality is also worth considering.

  • @ai_handbook
    @ai_handbook Місяць тому

    Very nicely explained, congrats mate! 👏👏👏

  • @zhalberd
    @zhalberd Місяць тому

    Thank you for this video. Please make a follow up on how to build this. 🙏🏻

  • @NevilHulspas
    @NevilHulspas Місяць тому +22

    Tried it. It's quite slow for answering simple queries. Also indexing about 20 pages in a PDF took about 300k tokens, which is still quite cheap with DeepSeek, but it seems a lot. Indexing also took like an hour or something. User interface is partly chinese, quite a bit of bugs. Seems unfinished.
    Answers that we're outputted were mostly correct though

    • @eshandas3645
      @eshandas3645 Місяць тому

      U used openai api key?

    • @조바이든-r6r
      @조바이든-r6r Місяць тому

      Did you tried gemini deep research

    • @ahmadzaimhilmi
      @ahmadzaimhilmi Місяць тому

      300k tokens for 20 pages is just too much.

    • @eshandas3645
      @eshandas3645 Місяць тому

      @@ahmadzaimhilmi thinking of using a local llama instead of deep seek

    • @AnuragVishwa
      @AnuragVishwa Місяць тому

      Which is the best RAG for json based raw data ?

  • @karthikbabu1844
    @karthikbabu1844 Місяць тому

    @engineerprompt, could you consider modifying the outline of your video thumbnails? Currently, when the bottom outline of a thumbnail is red, it often gives the impression that the video has already been watched. This can lead to confusion for viewers who might skip over new content, thinking they’ve seen it before.
    Great work so far!!

    • @engineerprompt
      @engineerprompt  Місяць тому

      Thanks for the feedback and makes total sense. Will look into it

  • @kai_s1985
    @kai_s1985 Місяць тому +2

    nice. wondering if we can use this on groq platform to speed things up?

  • @rishavranaut7651
    @rishavranaut7651 Місяць тому +2

    All these things still in infancy stage to be used in real use cases, ..will take time to match up with traditional RAG

    • @engineerprompt
      @engineerprompt  Місяць тому

      I agree! I am happy to see the progress and the research in retrieval. Retrieval is still one of the most useful applications of the LLMs.

  • @maniecronje
    @maniecronje Місяць тому

    Thank you for sharing great tutorial …reading from others comments it looks like the model is slow with responses which prevented me trying it but nevertheless informative nonetheless thanks 🙏

  • @alprbgt
    @alprbgt Місяць тому

    Hi, First of all, Thanks for your good video. It's knowledgeble video. I tried what you do in the video on my computer, but when I create a task, I always getting a vectorization connection error. I'm using OpenAI API and it's embedding model. My question, is there any document or any video on platforms about that error?

  • @patruff
    @patruff Місяць тому +11

    Not to be confused with CAG

    • @coom07
      @coom07 Місяць тому

      I got confused 4 a second. Thanks buddy❤

    • @andydataguy
      @andydataguy Місяць тому

      🤣🤣

  • @u007james
    @u007james Місяць тому +8

    how does continual update works with kag

    • @xt3708
      @xt3708 Місяць тому +3

      Same q!!

  • @Alex-rg1rz
    @Alex-rg1rz Місяць тому

    very interesting! thanks

  • @sanjaybhatikar
    @sanjaybhatikar Місяць тому

    It is the ReAct prompting with graph backend

  • @beppemerlino
    @beppemerlino Місяць тому

    amazing!

  • @matiasm.3124
    @matiasm.3124 Місяць тому +1

    Nice can you do the one in python using all local services/llm please.

  • @tirushv9681
    @tirushv9681 Місяць тому +1

    What do you think about Colpali vs KAG? which is the best?

    • @tirushv9681
      @tirushv9681 Місяць тому

      I feel Colpali performs well with accuracy and cost effective. Because image retrieval is from Colpali and we host it. Thoughts?

    • @MrAhsan99
      @MrAhsan99 Місяць тому

      @@tirushv9681 Is there anyone providing the api for colpali usage? I want to use this in my app. Want to give access via api to limited users/

  • @anjinsama59
    @anjinsama59 Місяць тому

    It seems now that openspg request Sign up with a phone number which is "Only Chinese Mainland (excluding Hong Kong, Macao and Taiwan) mobile phone number registration" at least when building the Docker image

  • @RodCoelho
    @RodCoelho Місяць тому

    How does it compare with light RAG? Is it Better?

  • @IamalwaysOK
    @IamalwaysOK Місяць тому

    Is there any way to use "Medical Information Extraction" instead of OpenIE?

  • @eshandas3645
    @eshandas3645 Місяць тому

    Can graph rag establish logical connection between them..

  • @DocRekd-fi2zk
    @DocRekd-fi2zk Місяць тому

    Is this based on rdf?

  • @david-journo
    @david-journo Місяць тому

    What about combining KAG and SPARQL ?

  • @abdshomad
    @abdshomad Місяць тому

    How is it compared to llm-graph-builder from neo4j?

    • @engineerprompt
      @engineerprompt  Місяць тому +1

      wasn't aware of it. Will need to check it out.

    • @definitelynotthefbi725
      @definitelynotthefbi725 Місяць тому +1

      ​@engineerprompt you really should, it's a fascinating framework

  • @surajjaiswal1371
    @surajjaiswal1371 Місяць тому +1

    What is better KAG or Agentic RAG?

    • @engineerprompt
      @engineerprompt  Місяць тому +1

      You can use KAG as a tool for an agent to do retrieval.

    • @surajjaiswal1371
      @surajjaiswal1371 Місяць тому

      @@engineerprompt Like the agent part in the Agentic RAG?

    • @engineerprompt
      @engineerprompt  Місяць тому

      @@surajjaiswal1371 exactly. RAG is just a tool that will be available to your agent.

    • @surajjaiswal1371
      @surajjaiswal1371 Місяць тому

      @@engineerprompt Okay, thanks!

  • @qadeer.ahmad123
    @qadeer.ahmad123 Місяць тому +1

    There is also CAG (Cache-Augmented Generation) much more faster.

  • @sanjaybhatikar
    @sanjaybhatikar Місяць тому +3

    Another day, another -ag 😂

  • @RohitSharma-uw2eh
    @RohitSharma-uw2eh Місяць тому

    Why hindi audio track is not available

    • @stefanmisic7405
      @stefanmisic7405 Місяць тому +1

      why would we need hindi audio??

    • @RohitSharma-uw2eh
      @RohitSharma-uw2eh Місяць тому

      @stefanmisic7405 because youtube add audio track feature is available.may we more enjoy like that mr beast videos

  • @ankitgadhvi
    @ankitgadhvi Місяць тому

    Very helpful tutorial for setup and good explanation on KAG systems however I used it and did not like the answers. it takes too long to give the answers and the answer are not good. also too much time needed to make embeddings from the chunks.

  • @NLPprompter
    @NLPprompter Місяць тому

    so this is like... um... uses Anthropic's contextual RAG + steroid Graph RAG?

  • @Sri_Harsha_Electronics_Guthik
    @Sri_Harsha_Electronics_Guthik Місяць тому

    need python version

  • @JoeCryptola-b1m
    @JoeCryptola-b1m Місяць тому

    No FAG? Fetched Augmented Generartion relax everyone. Uses internet fetch over docs that's why it's FAG not RAG

  • @JNET_Reloaded
    @JNET_Reloaded Місяць тому

    i notice it tries to use mysql i already have mysql and that default port ofc is in use this is bs! use sql lite or something like a file thats separate from a server and dont require ports ffs!

  • @JNET_Reloaded
    @JNET_Reloaded Місяць тому +1

    I would NOT reccomend this bs at all! terrable!

    • @matiasm.3124
      @matiasm.3124 Місяць тому

      @@JNET_Reloaded why ??? Explain

    • @iftekharshaikh7222
      @iftekharshaikh7222 Місяць тому +3

      @@matiasm.3124 they have basically being using SQL in the backend and the biggest problem with the Graph based approach is that it adds an LLM in between that increases cost and decreases speed in the indexing process and on top of that they are using 2 LLMs for indexing so double the cost and slower the speed also in the retrieval process they have query decomposition using NLP and stuff the chunk retrieval process is slower than usual as they are also using an LLM in the middle so this approach might get you the best results and relevance uptill now but it will come with high costs and slower speed as another person said that yiu have to wait 30 seconds for a single answer that too was not complicated so there is no practical use for this

  • @SonGoku-pc7jl
    @SonGoku-pc7jl Місяць тому

    kag waw! i try project with mypufd4llm, is posible combination with this?