Excellent presentation. Neo4j has done a great job of integrating vector embeddings with knowledge graphs. The neo4j LLM Knowledge Graph Builder for extracting entities and relationships from video transcripts, Wikipedia articles, and pdf format files is impressive. It also merges the extracted knowledge into a graph structure, and then provides an interface to query the LLM about knowledge in the graph.
Very useful info -- just what I was looking for as I'm researching this stuff. Also your delivery is great! Entertaining, yet you explain things very clearly. Subscribed.
Hi it would be great if you could please make longer videos explaining how you did each of these transition for example entity extraction, relationship extraction and so on and then how you did the neo4j integration. Maybe you can make a short video like this out of the original video to attract customers while the long video would still serve as a promising directions for the developer/researcher. I love the output produced from your system, but there’s no way to reproduce what you are doing. Reproducibility is a major concern in KGC.
Thanks for the feedback on producing deep dives into those topics. it's too much to cover in an applied video like this, but could be a good topic for a future video. In terms of reproducibility, you should be able to reproduce any of the examples in the video using the linked github project repo in the description.
Once the entities are extracted, can the model then be prompted to write a graph query (that could then be executed)? I’m thinking in particular of the “knowing A is B but not B is A” problem, such as when you ask an LLM “who is Mary Pfeiffer’s son?” and it does not say “Tom Cruise” but can answer “who is Tom Cruise’s mother?” just fine?
omg I was trying to hold it together at the introduction when I saw the large soup bowl but I broke down when it said "whatever, y'all get replaced by AI in 5 yrs" 😂 Great video overall. I hope I am right about the size of the soup bowl and you are not 3 feet tall.😅
Hey. Thx for the awesome content. Would it be possible to actually show a fully working large scale graph (like proper prod scale thing), and also discuss pros and cons of the approach, and when did you find KGs working well and not so well? The reason I asked is that my own KG experiments worked perfectly for me at smaller scale, but then the speed was so slow, that it was killing my M2 mac all together.
Hi, could you convert complex PDF documents (with graphics and tables) into an easily readable text format, such as Markdown? The input file would be a PDF and the output file would be a text file (.txt).
Maybe the pages have to be turned into images using Poppler and you could use an LLM that allows image inputs like GPT4Vision and Claude3 along with Function Calling to get the entities, objects and relations.
Unlike OpenAI, Diffbot does not train on your API inputs. Also, we offer on-premises solutions for enterprise customers that are self-hosted. Diffbot's services analyzes text and extracts entities and their relationships (aka a knowledge graph).
Excellent presentation. Neo4j has done a great job of integrating vector embeddings with knowledge graphs. The neo4j LLM Knowledge Graph Builder for extracting entities and relationships from video transcripts, Wikipedia articles, and pdf format files is impressive. It also merges the extracted knowledge into a graph structure, and then provides an interface to query the LLM about knowledge in the graph.
The neo4j graph builder (llm-graph-builder.neo4jlabs.com/) is indeed awesome! try selecting the "Diffbot" option in the generate graph dropdown
Amazing video really. and jokes are definitely necessary. 🤓
Very useful info -- just what I was looking for as I'm researching this stuff. Also your delivery is great! Entertaining, yet you explain things very clearly. Subscribed.
This is super cool! Thanks for sharing the detailed breakdown, super helpful. WIshing DiffBot the best as well
Thanks for the knowledge (graph) sharing :)
Thank you! This is exactly what I needed!
Awesome indeed 😍Thank you!
You rock, this was great!
Hi it would be great if you could please make longer videos explaining how you did each of these transition for example entity extraction, relationship extraction and so on and then how you did the neo4j integration. Maybe you can make a short video like this out of the original video to attract customers while the long video would still serve as a promising directions for the developer/researcher. I love the output produced from your system, but there’s no way to reproduce what you are doing. Reproducibility is a major concern in KGC.
Thanks for the feedback on producing deep dives into those topics. it's too much to cover in an applied video like this, but could be a good topic for a future video. In terms of reproducibility, you should be able to reproduce any of the examples in the video using the linked github project repo in the description.
Once the entities are extracted, can the model then be prompted to write a graph query (that could then be executed)? I’m thinking in particular of the “knowing A is B but not B is A” problem, such as when you ask an LLM “who is Mary Pfeiffer’s son?” and it does not say “Tom Cruise” but can answer “who is Tom Cruise’s mother?” just fine?
Yes! There are many people working on text-to-graph query language and that is a great motivating example for how to overcome language modeling bias
More like this please!
omg I was trying to hold it together at the introduction when I saw the large soup bowl but I broke down when it said "whatever, y'all get replaced by AI in 5 yrs" 😂 Great video overall. I hope I am right about the size of the soup bowl and you are not 3 feet tall.😅
LeanChen, if you had to rate the similiarity of groups of words, how would you do it? A bit like are these two classes similar?
Hey. Thx for the awesome content. Would it be possible to actually show a fully working large scale graph (like proper prod scale thing), and also discuss pros and cons of the approach, and when did you find KGs working well and not so well? The reason I asked is that my own KG experiments worked perfectly for me at smaller scale, but then the speed was so slow, that it was killing my M2 mac all together.
Great idea for a future video. For trying out a production version of a 10B node graph built from the entire public web, try out Diffbot!
I appreciate the meme cup 🎉😂
How does this compare to Graph RAG by Microsoft? What are their differences and similarities?
how does this compare to llamaindex property graph ?
Hi, could you convert complex PDF documents (with graphics and tables) into an easily readable text format, such as Markdown? The input file would be a PDF and the output file would be a text file (.txt).
Maybe the pages have to be turned into images using Poppler and you could use an LLM that allows image inputs like GPT4Vision and Claude3 along with Function Calling to get the entities, objects and relations.
would it work w/ ollama / local models?
DiffBot - sound not so good for privacy.
That is the role of the DiffBot? Entities extraction?
Are there any selfhosted replacements?
Unlike OpenAI, Diffbot does not train on your API inputs. Also, we offer on-premises solutions for enterprise customers that are self-hosted. Diffbot's services analyzes text and extracts entities and their relationships (aka a knowledge graph).
Awesome!
fugazi KG- this is only simple LPG (Label Property Graph)
Graphrag is so hot right now.
Why is the presenter speaking like a robot?
Bc she probably is AI-generated
be my teacher,and advisor please connect
She speaks in a very slurred way, I can barely understand her without turning subtitles on.
I can hear her fine. No issues.
She speaks completely fine.
too much information on the screen at once ... really painful to follow ... be more sober and straight to the point. Jokes are not necessary
Thanks for the feedback!
Why so serious? 😅