GraphRAG: LLM-Derived Knowledge Graphs for RAG

Поділитися
Вставка
  • Опубліковано 23 лис 2024

КОМЕНТАРІ • 142

  • @alexchaomander
    @alexchaomander  6 місяців тому +32

    What scenarios do you see GraphRAG being useful for?
    UPDATE: GraphRAG is now open source! Check out the release announcement video here: ua-cam.com/video/dsesHoTXyk0/v-deo.html

    • @jtjames79
      @jtjames79 6 місяців тому +10

      Using GraphRAG to make GraphRAGs.
      Because AI should be able to go down the rabbit hole.

    • @alexanderroodt5052
      @alexanderroodt5052 6 місяців тому +6

      Profiling people

    • @Sergio-rq2mm
      @Sergio-rq2mm 6 місяців тому +10

      Any where, where relationships are important. Abstract associations between data sets, perhaps laws, policies, etc, things that are very narrative driven, such as stories, etc. Nontypical datasets basically.

    • @alexanderroodt5052
      @alexanderroodt5052 6 місяців тому +1

      @@Sergio-rq2mm I choose to go the 1984 route

    • @ktbumjun
      @ktbumjun 6 місяців тому +5

      Bible study

  • @alexanderbrown-dg3sy
    @alexanderbrown-dg3sy 6 місяців тому +41

    This is basically causal grounding. We figure semantic symbolic reasoning, from an architectural perspective. Add a powerful model…something very compelling AGI-like would be the result I would assume(plus mcts sampling lol). Causal grounding is huge hole in current models.
    This is dope research. Kudos.

  • @BhaswataChoudhury
    @BhaswataChoudhury 6 місяців тому +42

    Looking forward to the code for this!

  • @jcourson8
    @jcourson8 6 місяців тому +14

    I've been doing work in the area of creating knowledge graphs for codebases. The nice thing about generating them for code (as opposed to text) is that you don't have to rely on LLM calls to recognize and generate relationships, but you can utilize language servers and language parsers for that.

    • @tomasb3191
      @tomasb3191 4 місяці тому +2

      that's interesting, what kind of insight can you get from the derived structure? I don't think code agents are leveraging language servers enough, it just looks like they only rag vector-search for context.

    • @reurbanise
      @reurbanise 4 місяці тому +2

      I'd love to hear more about this. Any code you can share?

  • @mvasa2582
    @mvasa2582 6 місяців тому +6

    While RAG is a good process for eliminating hallucinations, GraphRAG makes the retrieved context richer with its relationship-building techniques. The expense is worth it. Is the result set then re-graphed, or will the same query twice be as expensive?

  • @peteredmonds1712
    @peteredmonds1712 6 місяців тому +19

    this was so well explained, nicely done. my first thoughts:
    1. i'd be curious to see benchmarks with cheaper LLMs. from my experience, even much smaller models like llama-3-8b can come close to gpt-4 in this use-case (entity extraction and relationships). a little fine-tuning could likely match or surpass gpt-4 for much cheaper.
    2. i wonder how this could be augmented with datasources which already have some concept of relationships, ie wikipedia, dictionaries, hypertext.

    • @mrrohitjadhav470
      @mrrohitjadhav470 6 місяців тому

      i was having thoughts🙂

    • @Rkcuddles
      @Rkcuddles 6 місяців тому

      GPT 4 not understanding these deep relationships is bar far the biggest bottleneck in me using it. This is super exciting

  • @lalamax3d
    @lalamax3d 6 місяців тому +8

    glad, i didn't skip this and watched video, thanks for sharing knowledge. seems very impressive.

  • @ChetanVashistth
    @ChetanVashistth 6 місяців тому +4

    This seems very powerful. Thanks for sharing it and explaining it well.

  • @andydataguy
    @andydataguy 6 місяців тому +5

    That final streamlit app was awesome!!

  • @Mrbeastifed
    @Mrbeastifed 6 місяців тому +2

    Is there an Open source implementation of this or how could I build it into my own app?

  • @JasonSun386
    @JasonSun386 6 місяців тому +2

    Seems like the video was incomplete. Is there another part

  • @mrstephanwehner
    @mrstephanwehner 6 місяців тому +4

    Is there no standard comparison approach? For example one could take academic literature reviews, collect their references, throw in some more, and ask the llm system. Compare the result with the original review. There might be summaries available in the accounting and legal world, that could be used also

    • @alexchaomander
      @alexchaomander  6 місяців тому +3

      Comparison is tough! It's another area of research we're heavily invested in. But I like the ideas that you're bringing up!

    • @sathyanarayanbalaji2971
      @sathyanarayanbalaji2971 6 місяців тому

      true that validation would be required to compare the result.

  • @sairajpednekar8049
    @sairajpednekar8049 6 місяців тому +5

    May I know the underlying technology used for hosting the graph database? Was it Cosmos db?

    • @nas8318
      @nas8318 6 місяців тому +1

      Likely neo4j

    • @alexchaomander
      @alexchaomander  6 місяців тому +6

      It's graph database agnostic! You can use your choice of Graph DB. The technique is general enough to support multiple

    • @HediBen-r7t
      @HediBen-r7t 6 місяців тому

      It's not about the datbase, it's about the methodlogy. RDF or PL graphs should both work

  • @lifedownunderse
    @lifedownunderse 6 місяців тому +2

    I really enjoyed this video! What tool did you use to visualise the POD cast graph?

  • @jasonjefferson6596
    @jasonjefferson6596 6 місяців тому +1

    Does the repeated term“regular RAG” refer to setups using vector databases?

  • @heterotic
    @heterotic 6 місяців тому +2

    How is this any different then Self Organizing Maps for RAG?

  • @Aditya_khedekar
    @Aditya_khedekar 6 місяців тому +2

    Hii, i am working on solving the same problem of vector search rag is not good. can you plz share the code a tutorial will be even great !!

  • @TomBielecki
    @TomBielecki 6 місяців тому

    I really like the addition of hierarchical agglomerative summarization, which gives holistic aanswers similar to RAPTOR RAG strategy but with the better data representation of knowledge graphs. I'll need to read the paper to understand if embeddings are used at all in this, and whether relationships are labelled or if they just have a strength value.

    • @UtsavChokshi
      @UtsavChokshi 4 місяці тому

      Relationships are not labelled but they have descriptions.

  • @filippomarino861
    @filippomarino861 6 місяців тому +1

    This could be a game-changer in both public and private-sector intelligence analysis (as I am sure you figured out.) Looking forward to additional info - but what about the private dataset's format? Is it vectorized? If so, can we assume that there are optimal and sub-optimal approaches? (IOW, is it fair to assume vectorization can significantly impact GraphRAG's performance?)

  • @ir43
    @ir43 3 місяці тому

    Are there any accelerators to convert a typical knowledge corpus of unstructured text to a knowledge graph conducive for graph rag? I understand we need to extract entities and figure out relationships but who does that work? An LLM?

  • @crioss6803
    @crioss6803 2 місяці тому

    Hi, how about the abstract ideas visualization in multidimensional space eg. some mathematical ideas and relations to other ideas - eg. in form of knowledge graph and/or deeper details research possibilities tools? Are there any such tools based on ai?

  • @energyexecs
    @energyexecs 4 місяці тому

    Thanks for the video. I can see a Use Case in my energy industry. Does GraphRAG work across all "modes" and "modalities"?

  • @GigaFro
    @GigaFro 6 місяців тому

    Excuse me if I’m wrong… listened to this while exercising… but the main issue explored here for each question was that questions like “what are the top themes?” Cannot be answered by the LLM with vanilla RAG. Is this correct?
    If so, then if context size grows large enough this will be less necessary right?
    Furthermore, by introducing a graph that has communities premised on topics/themes or whatever u decide, doesn’t that reduce the degrees of freedom of your system?

  • @mohamedkarim-p7j
    @mohamedkarim-p7j 11 днів тому +1

    Thank for sharing👍

  • @vishwakneelamegam9479
    @vishwakneelamegam9479 5 місяців тому

    Is there a way to retrieve specific area of the graph than providing total graph to the LLM.

  • @escanoxiao6871
    @escanoxiao6871 6 місяців тому

    fabulous work! wondering how long it takes to form a whole vector db and plus how many tokens will it take?

  • @Cedric_0
    @Cedric_0 Місяць тому

    Any upadte on that streamlit code , that would be helpful thanks

  • @knutoletube
    @knutoletube 6 місяців тому +1

    Is the rest of this conversation available somewhere, @alexchaomander?

  • @phillipmaire8637
    @phillipmaire8637 5 місяців тому

    Would love the opportunity to contribute to this project, super interesting.
    How easy is it to update existing knowledge graphs periodically when new data comes in? Is there a “reindexing” cost?

  • @knaz7468
    @knaz7468 5 місяців тому +1

    Run this on the Lex Fridman podcast library!

  • @olegpopov3180
    @olegpopov3180 6 місяців тому +2

    What is technology stack for that?

  • @En1Gm4A
    @En1Gm4A 6 місяців тому +54

    pls provide the code

    • @alexchaomander
      @alexchaomander  6 місяців тому +25

      Code will be shared soon!

    • @SamuelJunghenn
      @SamuelJunghenn 6 місяців тому

      +1 🙏

    • @En1Gm4A
      @En1Gm4A 6 місяців тому +2

      @@alexchaomander Great! I have signed up for your newsletter. Will you inform about the code release there?

    • @Lutz1985
      @Lutz1985 6 місяців тому

      le dot

    • @bejn5619
      @bejn5619 6 місяців тому

      +1

  • @mhkk1122
    @mhkk1122 5 місяців тому

    I am waiting eagerly for the code of this paper.

  • @pablof3326
    @pablof3326 6 місяців тому

    Great work! I was thinking to use a system like this to build the memory of an AI companion as it talks to the user. So in this case the knowledge graph will start empty and grow get built dynamically with every conversation. Do you see this as a good use case for GraphRAG?

  • @joydeepbhattacharjee3849
    @joydeepbhattacharjee3849 3 місяці тому

    Very nice presentation and explanation

  • @mhkk1122
    @mhkk1122 5 місяців тому

    This approach is really good but don't you think that extracting entities and then making relationships between extracted entities is an expensive operation if we use GPT4 or Gemini?

    • @j.k.priyadharshini9753
      @j.k.priyadharshini9753 5 місяців тому

      I thought the same. Using knowledge graph is super but how we going to create it with less computation resources and less cost?

  • @PsyoPhlux
    @PsyoPhlux 4 місяці тому +2

    Pretty soon, everyone will be graphragging their podcasts. Jre will be neat.

  • @MyRandomnezz
    @MyRandomnezz 5 місяців тому

    Can you provide the code for this? Would be amazing!

  • @DefenderX
    @DefenderX 6 місяців тому

    Great, this is something I also thought about when AI had difficulty finding relevant information a while back.
    Basically have filters to determine how the AI will maneuver the training data depending on what is prompted and relevance.
    This is something I thought about after reading a paper on the discovery of a new hybrid braincell type that acted as a trigger that could turn on and off pathways.
    So the context in the prompt is what's important. Because that decides which tags in the training data should be turned on and off.
    Which in the end will give you a unique pathway for the AI to retrieve data.

    • @DefenderX
      @DefenderX 6 місяців тому

      Also, the next step would create overarching filters between several AI agents.
      After you have all this, the next step is for AI to implement statistics in its reasoning.

  • @sandeepsasikumar701
    @sandeepsasikumar701 5 місяців тому +2

    Is the code available?

  • @Rkcuddles
    @Rkcuddles 6 місяців тому

    Please let me play with this! Impressive work !

  • @hjl1045
    @hjl1045 6 місяців тому

    When will it be open sourced? :)

  • @ysy69
    @ysy69 4 місяці тому

    This is powerful!

  • @anthonyanalytics
    @anthonyanalytics 4 місяці тому

    This is amazing!

  • @SDAravind
    @SDAravind 5 місяців тому

    Whats the database used?

  • @ghostwhowalks2324
    @ghostwhowalks2324 6 місяців тому

    This is just brilliant

  • @Cedric_0
    @Cedric_0 3 місяці тому

    Streamlit code would be great, thanks

  • @NobleCaveman
    @NobleCaveman 6 місяців тому

    Would be a great tool for rapid and more reliable meta analysis

  • @FitoreKelmendi-fm1tg
    @FitoreKelmendi-fm1tg 5 місяців тому

    Does chatgpt (paid version) use graph rag?

  • @pabloe1802
    @pabloe1802 6 місяців тому

    To understand semantic search first you need to understand how HNSW works, then you realice no wonder it dosent work. I ended up building a datastructure to combine vector search and entities

  • @lesptitsoiseaux
    @lesptitsoiseaux 4 місяці тому

    Its a month later, where's the code you promised? Please?

  • @dhirajkhanna-thebeardedguy
    @dhirajkhanna-thebeardedguy 6 місяців тому

    This is outstanding stuff!

  • @Thrashmetalman
    @Thrashmetalman 6 місяців тому

    is there source code anywhere for this?

  • @ABG1788
    @ABG1788 5 місяців тому

    I don't understand. Why do we need GraphRag, when an LLM can summarise the text and find relationships ?

    • @mrpocock
      @mrpocock 5 місяців тому

      Knowledge graphs have a ton of formal methods to work with them. If you can get it into RDF then you can now use all the RDF tooling, or analyse it in cytoscape, or whatever.

  • @Sarmoung-Biblioteca
    @Sarmoung-Biblioteca 6 місяців тому

    GraphRAG Perfect !

  • @malikanaser8251
    @malikanaser8251 6 місяців тому

    Hi, are you going to share the code?

  • @RickySupriyadi
    @RickySupriyadi 6 місяців тому +1

    oh hey that's obsidian note style of note making
    it is interesting AI actually can remember better with the help of zettelkasten like human do!?
    can't wait until japan researcher conclude their research using chemical reactions in tube to emulate emotions, so machine can felt emotions through chemical reactions, like human do.... to me emotional are also the best way to learn and remembering things.

    • @RickySupriyadi
      @RickySupriyadi 6 місяців тому

      so what if... instead of tube of chemical reactions...
      important informations and often asked questions had an emotional cue graph to create some kind of important profiling so that profile will serve as a mark whenever AI is the expert in that field (strong retrieval in specific field leading for future of MoE)

  • @adilgun2775
    @adilgun2775 2 місяці тому

    great research topics but as an hands-on nlp engineer on ner boosted knowledge graphs and LLMs , my experiences say too naive to believe that it would work in productions systems

  • @tacticalgaryvrgamer8913
    @tacticalgaryvrgamer8913 5 місяців тому

    I assume it's open source because why would someone pay to have gpt4 parse and organize their data. Takes 2 seconds to roll your own.

  • @MahmoudAtef
    @MahmoudAtef 6 місяців тому +1

    But knowledge graphs are very slow to query. I wonder if we can encode those graphs in the gpt model by building graph transformers.

    • @damianlewis7550
      @damianlewis7550 6 місяців тому +3

      I don’t think that’s the case. Optimized graph query engines can return results in milliseconds e.g. WikiMedia, Google etc. at a fraction of the computational cost of an LLM.
      The reason that GraphRAG is slow-ish is because the LLMs are slow.

    • @MrDonald911
      @MrDonald911 6 місяців тому +4

      Google, Facebook, and Linkedin all use graph databases, it's actually much faster than relational DBs

    • @nas8318
      @nas8318 6 місяців тому +2

      Slower than LLMs?

  • @youcaio
    @youcaio 4 місяці тому

    è bellissimo!

  • @nickfleming3719
    @nickfleming3719 5 місяців тому

    Okay... we know graph rag is good. duh. How is it implemented, how do you feed it to the LLM, how do you store the data

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w 6 місяців тому +4

    but don't you lose information in the process of making a knowledge graph, given how only a subset of the textual information is extracted and retained in the KG?

    • @computerrockstar2369
      @computerrockstar2369 6 місяців тому

      I don't think the LLM really needs the graph to make any decisions. Its more valuable for human users to find related information

    • @HediBen-r7t
      @HediBen-r7t 6 місяців тому

      You can use ETL to build your knowledge graph by yourself from RDMSs, then you will not loose information

  • @Sri_Harsha_Electronics_Guthik
    @Sri_Harsha_Electronics_Guthik 6 місяців тому

    implementations?

  • @vcool
    @vcool 4 місяці тому +1

    Admit it: y'all built graphrag first for use by the CIA. This is not a joke.

  • @lanc3carr
    @lanc3carr 6 місяців тому

    Police, FBI, CIA, etc... investigations (CSI AI)

  • @bicepjai
    @bicepjai 4 місяці тому +1

    Why do we need other faces on screen ? Hope they know they are just distractions :)

    • @nikharmalik6090
      @nikharmalik6090 4 місяці тому

      And I hope you know that you can zoom in to the screen to not see them and that it's always better not to say anything if you don't have anything nice or useful to say 😊

  • @Walczyk
    @Walczyk 6 місяців тому

    What's a rag

    • @IlyaDenisov
      @IlyaDenisov 6 місяців тому

      Retrieval Augmented Generation (use that as an input to your favourite search engine or AI companion)

  • @Ian_Arden
    @Ian_Arden 4 місяці тому +1

    Where did you get the body for this? This whole text is taken from some chief Russian propaganda bureau 😅

  • @RameshK-c1w
    @RameshK-c1w 6 місяців тому

    American princess Google Plex SEO Sandra Mitra watching.....

  • @historia_tego_swetra
    @historia_tego_swetra 4 місяці тому

    first of all, there is no MOVEMENT but state sponsored Russian proxies like the Yemenis.
    a very wrong choice of a dataset.
    second thing - there is NO NOVOROSSIA

  • @cyberslot
    @cyberslot Місяць тому

    You could save yourselves the political/propaganda element.
    From all uncounted number of articles available, the choice of this particular topic is more than flashy. It's ridiculous for techie persons who are expected to be smart in general...

  • @ross9263
    @ross9263 6 місяців тому +2

    The content is very political..

    • @somjrgebn
      @somjrgebn 5 місяців тому +1

      Haha, and skewed... Crickets for Gaza... but Odessa is worth mentioning?
      This is why it's best to avoid politics when we're trying to stay on task, especially when dealing with tech that's literally forming and pruning knowledge graphs based on topics/themes...