I believe you are the hero every AI engineer needs. Unlike most UA-camrs who copy and paste code from documentation, you address the real problems AI engineers face.
Very helpful, Dave! Many companies try to use naive chunking because there are so many examples on the web, UA-cam videos, etc. You gave us a very good way to do smarter chunking and get more useful results. This is the future for RAG use cases.
Hi Dave, Thanks for the awesome content, a client come to me for a RAG solution, he have a library of hundreds of thousands of pages (about 60 Giga), simplest rag techniques doesn't seem to work for this case, come up to a solution using hybrid retriever and a reranker using llama-index, the results was good but not perfect, if were you how will you tackle this problem?
we are working on a solution for this that can be white-labeled on release! does your client has an API endpoint or some kind o bucket containing all the files ? it really depends in what formats the data comes. If it its just text than you can use a hybrid-approach with semantic chunking, parent-document retrieval or other meta-data filtering techniques. The main point of importance is to make sure the data is pre-processed and cleaned before being chunked an embedded. Entity extraction is expensive but can be very helpful. A second best option is to extract meta-data. One is used for semantic extraction (entity) and the other for additional filtering. GraphRAG is the best solution, using entities, but it costs a massive amount of resources & development time making it only accessible to enterprise clients (10-50k +).
Hey Dave , Really Nice Video . I was wondering if I could help you with more High Quality and engaging editing with maintaining a brand colour to your youtube channel which can help you to get more engagement in your videos and Build your Unique Personal Brand . Pls lmk what do you think ?
Will try this with textaract. For my use case I am just sending a csv ( of an excel ) and its working but I think that is not a systematic, luck proof way. Do you think rag approach will be better, less prone to context, structure related hallucinations ?
GraphRAG is a more powerful solution than this baseline RAG. In GraphRAG, the data is stored in the graph with entities and relationships and also doing community summaries in detail which excels in retrieval flow. For eg: questions like "Did company underperform in Q4 vs Q3?" This kind of question would be difficult to answer using Baseline-RAG which can be answered easily using GraphRAG
@@sahiljain9376 you can enhance RAG with agentic frameworks to allow these questions e.g. an SQL Agent with meta-data filtering. I love graphRAG but its a.) super expensive since entity extraction requires a ton of LLM calls b.) takes allot of time to set-up the graph, c.) has additional challenges to be overcome before it can really be used for non-enterprise.
@@sahiljain9376 I was unaware of GraphRAG, and it looks really interesting thanks. It looks like it's beyond my skill level now, but hopefully MS integrates it into Azure soon
I believe you are the hero every AI engineer needs. Unlike most UA-camrs who copy and paste code from documentation, you address the real problems AI engineers face.
Appreciate that!
So good at simplifying concepts in these tutorials. Loved this Dave!
Thanks Brock 🙏🏻
Fucking awesome. Best AI youtuber for me. Keep up the good work, Dave!
Very helpful, Dave! Many companies try to use naive chunking because there are so many examples on the web, UA-cam videos, etc. You gave us a very good way to do smarter chunking and get more useful results. This is the future for RAG use cases.
Thanks Clarke!
Your video is detailed and very helpful, thank you for these type of techniques.
Thanks a lot for sharing this knowledge, it's really useful!
Good stuff! We use the exact same technique with markdown-based chunking and extra metadata for the chunks. Works really well!
I think this is currently the best approach for RAG.
great - thank you for sharing:) Please explore the topic more - )
Super helpful. Can you please make a tutorial on how to use AWS Textract too?
Hi Dave, Thanks for the awesome content, a client come to me for a RAG solution, he have a library of hundreds of thousands of pages (about 60 Giga), simplest rag techniques doesn't seem to work for this case, come up to a solution using hybrid retriever and a reranker using llama-index, the results was good but not perfect, if were you how will you tackle this problem?
we are working on a solution for this that can be white-labeled on release! does your client has an API endpoint or some kind o bucket containing all the files ? it really depends in what formats the data comes. If it its just text than you can use a hybrid-approach with semantic chunking, parent-document retrieval or other meta-data filtering techniques. The main point of importance is to make sure the data is pre-processed and cleaned before being chunked an embedded. Entity extraction is expensive but can be very helpful. A second best option is to extract meta-data. One is used for semantic extraction (entity) and the other for additional filtering.
GraphRAG is the best solution, using entities, but it costs a massive amount of resources & development time making it only accessible to enterprise clients (10-50k +).
Hey Dave , Really Nice Video . I was wondering if I could help you with more High Quality and engaging editing with maintaining a brand colour to your youtube channel which can help you to get more engagement in your videos and Build your Unique Personal Brand . Pls lmk what do you think ?
Thanks I'm a newbie and your videos helped get me started. Can you please also share pdf_ingester?
Enjoyed this. Thank you.
Have you tried passing the PDF to Jina Reader API? The Markdown output is quite clean too! (but it's only usable for public documents)
Could you do a GraphRAG tutorial?
awesome video but where can we find the "from config.settings import get_settings" ?
Will try this with textaract. For my use case I am just sending a csv ( of an excel ) and its working but I think that is not a systematic, luck proof way. Do you think rag approach will be better, less prone to context, structure related hallucinations ?
could you please share the pdf_ingester code too. I would like to play around with it
Does it work with scanned pdf docs?
Yes!
This seems similar to GraphRAG. What is the difference?
GraphRAG is a more powerful solution than this baseline RAG. In GraphRAG, the data is stored in the graph with entities and relationships and also doing community summaries in detail which excels in retrieval flow. For eg: questions like "Did company underperform in Q4 vs Q3?" This kind of question would be difficult to answer using Baseline-RAG which can be answered easily using GraphRAG
@@sahiljain9376 you can enhance RAG with agentic frameworks to allow these questions e.g. an SQL Agent with meta-data filtering. I love graphRAG but its a.) super expensive since entity extraction requires a ton of LLM calls b.) takes allot of time to set-up the graph, c.) has additional challenges to be overcome before it can really be used for non-enterprise.
@@sahiljain9376 I was unaware of GraphRAG, and it looks really interesting thanks. It looks like it's beyond my skill level now, but hopefully MS integrates it into Azure soon