Data Centric
Data Centric
  • 50
  • 282 748
AI Agent for Deep Web Search with o1 Preview (Open-Source)
In this video, we delve into Jar3d, an open-source AI agent developed with LangGraph, designed to surpass traditional search engines by conducting deep, long-running web research and information synthesis. Jar3d stands as an open-source alternative to Perplexity, capable of performing complex tasks like market research, product sourcing, newsletter writing, and grant finding.
We demonstrate the power of Jar3d using OpenAI's new o1-preview model, showcasing its enhanced capabilities in deep web search and research tasks. The video also highlights new features of Jar3d, including feedback mechanisms and task tracking. Additionally, we discuss the advantages of using o1-preview over Claude 3.5 Sonnet and explore potential improvements and future developments for Jar3d.
I provide consultancy services, book a discovery call here: www.brainqub3.com/book-online
Sponsor the development of Jar3d: github.com/sponsors/brainqub3
Register your interest in the AI Engineering Take-off course: www.data-centric-solutions.com/course
GitHub repo for Jar3d: github.com/brainqub3/meta_expert
Meta Prompting Research Paper: arxiv.black/pdf/2401.12954
Chapters: Intro: 00:00
Jar3d at a Glance: 01:59
Setting up Jar3d with Docker: 07:34
Jar3d Internet search with o1-preview: 12:55
Final thoughts on o1-preview: 42:33
Переглядів: 6 029

Відео

Neo4j Knowledge Graphs for "Smarter" AI-Agent Search
Переглядів 5 тис.Місяць тому
Jar3d is an open-source AI agent that leverages sophisticated AI engineering techniques, including meta-prompting, RAG, and advanced prompt engineering orchestrated by LangGraph. Jar3d can perform long-running, research-intensive tasks that require information from the internet. This video gives launches a new feature that incorporates Neo4j knowledge graphs into Jar3d’s context to improve the ...
How this AI Agent Uses LangGraph & Prompt Engineering to Challenge Perplexity (Deep Dive)
Переглядів 9 тис.Місяць тому
How this AI Agent Uses LangGraph & Prompt Engineering to Challenge Perplexity (Deep Dive) Jar3d is an open-source AI agent that rivals Perplexity and other search agents in conducting deep research tasks. This agent leverages advanced AI engineering techniques, including: Meta-prompting, Retrieval-Augmented Generation (RAG), Sophisticated prompt engineering, and LangGraph for orchestration. The...
AI Agent | Perplexity Alternative Built with LangGraph & Advanced Prompt Engineering (Demo)
Переглядів 5 тис.Місяць тому
Jar3d is an open-source AI agent that leverages sophisticated AI engineering techniques, including meta-prompting, RAG, and advanced prompt engineering orchestrated by LangGraph. Jar3d can perform long-running, research-intensive tasks that require information from the internet. This video demonstrates how Jar3d works with the Llama 3.1 70B model, effectively creating an open-source version of ...
Dynamic AI Agents with LangGraph, Prompt Engineering Enhancements + RAG
Переглядів 9 тис.2 місяці тому
Combining prompt-engineering techniques such as chain-of-reasoning and meta-prompting with Retrieval-Augmented Generation (RAG) on the fly has enabled me to develop a powerful agent for long-running, research-intensive tasks. Jar3d has internet access and significantly enhances tasks like creating newsletters, writing literature reviews, planning holidays, and other research-intensive activitie...
A Prompt Engineering Trick for Building "High-level" AI Agents
Переглядів 11 тис.2 місяці тому
We explore a handy prompt engineering technique designed to help you build more flexible and powerful AI agents. We'll look at how meta-prompting works under the bonnet, and talk through an implementation of the prompting framework with a web search AI agent. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your interest in the AI Engineering Take-off course: www.data...
AI Agents: Why They're Not as Intelligent as You Think
Переглядів 4,2 тис.3 місяці тому
I will be pushing AI agents to their absolute limits by testing the most powerful models available today against a computer chess model. The test reveals how effective LLM-powered AI agents are at planning and highlights some limitations that you must be aware of if you are building with AI agents. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your interest in the ...
Building AI Agents from Scratch, Simplified
Переглядів 24 тис.3 місяці тому
If you’ve always wondered how AI agents work under the hood, this video is for you. I’ll be revealing the mechanics of AI agents, and given you a basic pattern you can use in your own projects to build any AI agent. You won’t require any intermediate libraries like LangChain or LlamaIndex for this. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your interest in the ...
LangGraph AI Agent Upgrade: Groq, Gemini, and Chainlit Front End
Переглядів 7 тис.3 місяці тому
This video follows on from my "LangGraph Made Easy" video. Join me as I walk you through the additional functionality and integrations (with Groq, Gemini, and Anthropic’s Claude) I've added, the bugs I've fixed, and the stylish front end I've created for my LangGraph custom AI agent, which searches the web like Perplexity. Need to develop some AI products? Let's chat: www.brainqub3.com/book-onl...
How I Build Local AI Agents with LangGraph & Ollama
Переглядів 7 тис.3 місяці тому
Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your interest in the AI Engineering Take-off course: www.data-centric-solutions.com/course Hands-on project (build a basic RAG app): www.educative.io/projects/build-an-llm-powered-wikipedia-chat-assistant-with-rag Stay updated on AI, Data Science, and Large Language Models by following me on Medium: medium.com/@johnadeo...
LangGraph Simplified: Master Custom AI Agent Creation
Переглядів 33 тис.3 місяці тому
Here’s a LangGraph tutorial that should put your mind at ease. There is significant interest in the LangGraph framework due to its customisability and integrations. However, many have found the core concepts to be quite complex. This video breaks down those core concepts in a friendly and digestible manner and provides you with a Python tutorial for a custom web search agent to help you concret...
Can Open Source AI Agents Beat Perplexity AI? Testing Codestral, GPT4o, and Mixtral
Переглядів 3 тис.3 місяці тому
This is the fourth and final video in a series where I test various open-source models with my custom web search AI agent to evaluate their performance. In this video, I benchmark the Mistral AI models Codestral 22B and Mixtral, as well as GPT4o, against Perplexity AI to see if my web search agent measures up. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your inte...
LLAMA 3 or Phi 3 AI Agent: Can they Beat Perplexity in Web Search?
Переглядів 2,1 тис.3 місяці тому
This is the third in a series of videos where I will be testing various open-source models with my custom web search agent to see how well they perform. In this video, I benchmark Llama 3 70b (instruction tuned by TenyxChat), gpt-3.5 turbo, and phi-3 medium against Perplexity AI to see if my web search agent measures up. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Registe...
Llama 3 70B Custom AI Agent: Better Than Perplexity AI?
Переглядів 4,2 тис.4 місяці тому
This is the second in a series of videos where I will be testing various open-source models with my custom web search agent to see how well they perform. In this video, I benchmark Llama 3 70b, hosted on Runpod, against Perplexity AI to see if my web search agent measures up. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your interest in the AI Engineering Take-off...
Can My Ollama Local WebSearch Agent (With Llama 3 8B) Beat Perplexity AI?
Переглядів 4 тис.4 місяці тому
This is the first in a series of videos where I will be testing various open-source models with my custom web search agent to see how well they perform. In this video, I benchmark Llama 3 8b, hosted locally, against Perplexity AI to see if my web search agent measures up. Need to develop some AI? Let's chat: www.brainqub3.com/book-online Register your interest in the AI Engineering Take-off cou...
Build Open Source "Perplexity" agent with Llama3 70b & Runpod - Works with Any Hugging Face LLM!
Переглядів 6 тис.4 місяці тому
Build Open Source "Perplexity" agent with Llama3 70b & Runpod - Works with Any Hugging Face LLM!
Build your own Local "Perplexity" with Ollama - Deep Dive
Переглядів 10 тис.4 місяці тому
Build your own Local "Perplexity" with Ollama - Deep Dive
Agency Swarm: Why It’s Better Than CrewAI & AutoGen
Переглядів 23 тис.4 місяці тому
Agency Swarm: Why It’s Better Than CrewAI & AutoGen
Forget CrewAI & AutoGen, Build CUSTOM AI Agents!
Переглядів 25 тис.4 місяці тому
Forget CrewAI & AutoGen, Build CUSTOM AI Agents!
Why I'm Staying Away from Crew AI: My Honest Opinion
Переглядів 28 тис.4 місяці тому
Why I'm Staying Away from Crew AI: My Honest Opinion
How to get LLaMa 3 UNCENSORED with Runpod & vLLM
Переглядів 3,7 тис.4 місяці тому
How to get LLaMa 3 UNCENSORED with Runpod & vLLM
Host Your Own Llama 3 Chatbot in Just 10 Minutes! with Runpod & vLLM
Переглядів 2,5 тис.5 місяців тому
Host Your Own Llama 3 Chatbot in Just 10 Minutes! with Runpod & vLLM
Mistral AI vs Open AI: Who REALLY Has The Best AI? Rap battle using Suno AI
Переглядів 5925 місяців тому
Mistral AI vs Open AI: Who REALLY Has The Best AI? Rap battle using Suno AI
WHY Retrieval Augmented Generation (RAG) is OVERRATED!
Переглядів 3 тис.5 місяців тому
WHY Retrieval Augmented Generation (RAG) is OVERRATED!
Is AutoGen just HYPE? Why I would not use AUTOGEN in a REAL use case, Yet
Переглядів 5 тис.7 місяців тому
Is AutoGen just HYPE? Why I would not use AUTOGEN in a REAL use case, Yet
Your GUIDE to Hugging Face, GPUs, OpenAI, LangChain + More in the LLM Ecosystem - Lecture 2
Переглядів 3,2 тис.7 місяців тому
Your GUIDE to Hugging Face, GPUs, OpenAI, LangChain More in the LLM Ecosystem - Lecture 2
Deploy Mixtral, QUICK Setup - Works with LangChain, AutoGen, Haystack & LlamaIndex
Переглядів 1,2 тис.8 місяців тому
Deploy Mixtral, QUICK Setup - Works with LangChain, AutoGen, Haystack & LlamaIndex
Building Chatbots with Hugging Face LLMs: 5 Expert Tips ft. Mistral
Переглядів 1,1 тис.8 місяців тому
Building Chatbots with Hugging Face LLMs: 5 Expert Tips ft. Mistral
CUSTOM RAG Pipelines & LLM Fine-Tuning: A GRADIENT Tutorial
Переглядів 9008 місяців тому
CUSTOM RAG Pipelines & LLM Fine-Tuning: A GRADIENT Tutorial
Zapier AI Gmail Automation: How to Automate Mundane Tasks and Save Hours
Переглядів 7269 місяців тому
Zapier AI Gmail Automation: How to Automate Mundane Tasks and Save Hours

КОМЕНТАРІ

  • @AbdallaMosa-w8j
    @AbdallaMosa-w8j 14 годин тому

    If you could choose one AI agent framework that you believe will become the industry standard, which one would it be?

  • @juanjosetorresvillarroel820
    @juanjosetorresvillarroel820 6 днів тому

    Thank you sir

  • @MecchaKakkoi
    @MecchaKakkoi 6 днів тому

    Great video! I would definitely be interested in seeing some evaluation methods. Cheers!

  • @1msirius
    @1msirius 8 днів тому

    I really love your videos <3

  • @xspydazx
    @xspydazx 8 днів тому

    very good , i was not actually sure how to compile up a docker : !!

  • @ChrisAlas
    @ChrisAlas 9 днів тому

    Dude is an ai video himself

  • @tubedelux9748
    @tubedelux9748 9 днів тому

    How dare you compare this to perplexity that a click away it literally took you an hour to explain how to use your tool

  • @jarad4621
    @jarad4621 9 днів тому

    Ive been away for a while, not by choice but excited to catch up on my favorite channel, this looks great so far. You are doing exactlt what i wanted to do am advanced long form output research agent so this is great wish i could sponser but im hoping i can in 2 to 3 months when things are going better

  • @EmminiX
    @EmminiX 9 днів тому

    WOW. glad you popped on my feed ! Question : is it possible to change the system prompt ?

  • @Karl-Asger
    @Karl-Asger 9 днів тому

    Great work

  • @mimameta
    @mimameta 10 днів тому

    Very cool, subscribed

  • @zacboyles1396
    @zacboyles1396 10 днів тому

    Adopting the Chainlit task list definitely enhances the responsive feel but given the slow speed of o1 I’d recommend generating a unique project id that you and the agent work on and use the webhook session restore to report status messages. Each message can have preset actions that replace your slash commands. That will enable you to offer feedback on the fly, instead of aggregating all the feedback into a single message. That slightly increases the complexity, more or less given the construction of your langgraph state classes. I cloned the repo last night, hopefully I’ll have a chance to look at the code and see if there’s any PR I could do.

    • @Data-Centric
      @Data-Centric 8 днів тому

      This sounds awesome. I appreciate this will be a lot of work but it will certainly enhance the user experience.

  • @CUCGC
    @CUCGC 10 днів тому

    It is awesome, I am going to contact you once I redo my project with the new concepts. I think I can use your help.

  • @mikew2883
    @mikew2883 10 днів тому

    Awesome work here! 👏 For those of us that are not Blessed with having OpenAI Tier-5 access, is it possible to offer an OpenRouter setting in this project so we can give this a try? 😁

  • @avi7278
    @avi7278 10 днів тому

    Anyone can get access to the o1-preview model via open router

  • @JNET_Reloaded
    @JNET_Reloaded 10 днів тому

    can we run this with local llm olama models?

  • @DonFrankAlvarez
    @DonFrankAlvarez 10 днів тому

    Thank you for your breakdown. I found especially useful to understand which model is better for what task in an agentic framework. I'm guessing o1 will also be quite expensive on top of the slower speed, so it should be reserved for mostly complex tasks where they need to follow steps to the T

  • @xspydazx
    @xspydazx 11 днів тому

    Lovely as usual : I think you need to make the decision on the embedding database and rag setup ! , is it over use ? are you using a hammer when a screwdriver is required ? As you know rag is an augmentation to creating addtional content for the model submission of a query : But it can also just be a simple knowledge base also : hece the rag itself can be a tool : ie a knowledge base for the model to search for infroamtion before doing a web search : So when a task is placed it can first search its knowledge base then search the web ! if required : So we should be thinking that we may need an additional service ! this is lets call it the web crawler service : SO as researcher you will know your topics that your interested in : so you should be able to put the model to work searching the internet for topics and creatig a mini collection of documents to be processed into the rag for use off line ! this will enable for local research !( SHORT TERM MEMORY) so when youo ask a question regarding topics you have rresearched they will be already held locally ! ( also you can add a date -Marker ! this will tell the model to seek updated knowledge if the current data is more then a month or two old ! ) ... hence blocking thr web search and keeping the work local ! this gives you the advantage of faster quering! , as you have a rag and it should be used as a knowledg base ie coolecting your browsing historys and merging them to the rag ! so also perhaps a mini browser in app to display pages ( perhaps converted to markdown ) ( aha a markdown browser automatically convverting webpages to markdown !!) ... this ienables you to save your pages after searching as markdown pages and even to a search dictionary (json ) and later fed to the rag or Fine tuned in to the model ! SO now the app becomes a true research tools ! So you should also add more rag tools to retrive various formatted documents to your UI ! SO ow you can have a seties of good tools to acess and manage your knowledge base and other for searching ! ie local or online ! or for knowledge building ! Or Report creation tools to display your rih information in the ui , or even DataBuilding ( corpus building Feature ) to build corpusses based on various searches in your reseach stack to produce docuemtns or database for fine tuninng into your model ! ( hence being able to utilize the rag as a Trasactional database as well as alocal knowldgebase enhancement ) we would exxpect that trained data can still be held in the rag but should be marked for archiving : as the modle should have the knowledge internally ! So on Rag optimization the model can remove these entrtyu from the rag as they are now in the pretrained stack ! IE THE LONG TERM MEMORY.... Obvioulsy your chat historys should always be optimizes by sumarization and also uploaded into the rag , before closing ! hence your WORKING MEMEORY become Uploaded in to your SHORT TERM MEMORY ! .... } So he we implement the full Memeory Stack Also ! this also enables for personalization of the model ! As local history and commuication and local knowledge is being optimized ito the rag : as well as being outputted in to a datset for training into the pretrain Stack: full complete pathway ! I specify again the most importsant thing is to train a model local model for Planning ! this really changed the model for me even after a year of training all these methodologes the modle seemed comparitive to larger models despite being 7b ( only can be run local for me ) < SO i think it was sucessfull ! but when i trained the model onn the react process and its clones as well as PLANNING ! the model Changed ! : So now when i used it for Mastro and aider etc as all the models i found now its planning really augmented its coding and tasks ! as well as its responses for normal nformation al querys !... I also implemented the repl tool ( runs notebook cells ) ... which also made the model always echeck the output before returning the code ! , Every thing changed .......Now its totally on point ! I think these were the missing components : As well as i changed my strategy : all of these think, action, observation etc these stages are all just prompts ! , SO a single self query tool is all you need and the pre wrtten prompts : these entry points are used to create agents ( or preloaded tools ( self-Query ) ) so the model can use the prompt as a tool ! LOL !!! << KEY ! So if i explain a type of execution strategy it can create the neccasaryu prompts ( for these tools ) ... Now a final Response can also be a tool , as the tool will format the response and even execute all code snippets ! As i added this to my input and output loops : so before i send code to the model , ( i used the markdown tag ) it will execute these snippits and pass the executions to the model ! so it will see the output from these codes and any output it sends to me will also be executed ! This interception also changed the model behavior ! to really being great with coding ! (python) ... Obvioulsly i also spend a lot of time collecting DOCSS from librarys ! and feeding these docs to the model ! <<< , So i do not use rag ! , but i understand the methodolgy and how it should be utilized for the full working memory, short term memory , and long term memory as i am also a Business inteligence Data Scientist ! (BSC/MA) ...( another day i will go over the data mining strategy ( web mining ) - ) I hope this hellp you to percive new directions for th app ! as you did enlighten me and force me to make the graph ( from scratch ) and design the chain! and begin to use state for every thing ! Which also Raised the BAR ! ... you did not convert me to RAG ! ... LOL !

  • @madhavpr
    @madhavpr 11 днів тому

    Hi @Data Centric. Do you plan to add a video on how you can make two (or more) agents talk to each other to fulfill a task from scratch? :)

  • @tsap1
    @tsap1 11 днів тому

    EIC not applicable to UK companies.

  • @cyberprotec
    @cyberprotec 11 днів тому

    I've been away for a while, and I'm thrilled to return and see the amazing progress and new features that have been implemented. It all makes perfect sense, and I’m excited to try this out. Thank you for your hard work!

  • @terminally_lazy
    @terminally_lazy 11 днів тому

    It's available on openrouter...

  • @randyblasik7066
    @randyblasik7066 11 днів тому

    Nice love your work.

  • @WadeInToAI
    @WadeInToAI 11 днів тому

    If you want access to openai's GPT o1 model via api and dont have the right tier you can go through openrouter and access it now.

  • @GNARGNARHEAD
    @GNARGNARHEAD 11 днів тому

    That sounds really interesting! I'll spin up a copy and give it a try. I've been meaning to explore integrating Neo4j with OpenAI, hould be fun to see how it goes. Thanks!

  • @luismonge8720
    @luismonge8720 11 днів тому

    Thank you for sharing this and explaining it in such detail. I get really excited when I see a new video of yours pop up on my subs tab. Curious, how big of an improvement do you think this is compared to just using gpt4o for all the agents?

  • @puneet1977
    @puneet1977 11 днів тому

    Love what you are up to with Jar3d.

    • @jarad4621
      @jarad4621 9 днів тому

      Best name for an app ever

  • @larissabjfd7u4tjirsdj
    @larissabjfd7u4tjirsdj 12 днів тому

    can i use this with ollama?

  • @nizammuhammed7586
    @nizammuhammed7586 12 днів тому

    Great video! You did an awesome job explaining the concepts. I just wanted to clarify one point regarding LangChain. It already includes libraries for GROQ and local models. For GROQ, we can use langchain_groq, and for any other local model, we can use: ChatOpenAI(api_key="ollama", model="llama3", base_url="localhost:11434/v1") Given this, I'm curious why you opted to create custom code under /models. Just want to make sure I'm fully understanding your approach.

  • @leopoldschmidt7085
    @leopoldschmidt7085 16 днів тому

    Question what Hardware do you use ?

  • @mahiaravaarava
    @mahiaravaarava 21 день тому

    Building AI agents from scratch involves designing algorithms that can learn and make decisions. Start with defining the problem, select appropriate machine learning models, and train them using relevant data. Simplify the process by breaking it into manageable steps and using frameworks that streamline development.

  • @Ken00001010
    @Ken00001010 24 дні тому

    Excellent explanation!

  • @amimaron1
    @amimaron1 25 днів тому

    Great video! Thanks for the well-structured code and clear explanations

  • @kennethreilly7648
    @kennethreilly7648 25 днів тому

    knowing what you know now about Langgraph, does your position change re: your suggestion/preference to build custom workflows from scratch? I recall you mentioning that preference in your CrewAI tutorial & you don't mention that at the end of your video here. Has Langgraph converted you to frameworks?

  • @anmolmittal9
    @anmolmittal9 26 днів тому

    Great Video!! I wanted to check how can we show these tool calls and other things as Chainlit steps?

  • @thoughtprovokingideas-j9t
    @thoughtprovokingideas-j9t 27 днів тому

    very informative video!

  • @brucehe9517
    @brucehe9517 27 днів тому

    re you planning to use Neo4j exclusively for your project, or do you still require another database for semantic search? I asked ChatGPT, and it seems that while Neo4j can handle semantic search, it may not be optimized for quick searches if the database becomes too large.

  • @efraijo
    @efraijo 27 днів тому

    I think it would be great if we/community could implement the ability to upload PDF documents for chunking and adding to the knowledge graph. I have not had any time to thinker on adding features but I've been playing with the system prompt and graph structure.

  • @SolidBuildersInc
    @SolidBuildersInc 27 днів тому

    Good Day Sir, I am interested in using a 70B model and at Timestamp 58:50 is where you discuss other models to your project , so is it possible to add new models without breaking the code? Appreciate any feedback.

  • @SixTimesNine
    @SixTimesNine 27 днів тому

    You don't reply to issues on GitHub. That +is+ necessary ... please :)

  • @xspydazx
    @xspydazx 28 днів тому

    very good ! ( but ... Think Should be a tool ) ( observe should be the return tool call ) ( action : should be the selected tool or action ) ( think : should be a tool to self query for a plan or methodology or how to porceed to the next step of a plan ? ) .... This way we can create Any methodology or chain of thoughts : if we make the components Tools themselfs .. such that : create a plan ( should be a tool ( which is a prompt , to query itself with a task , so given a task as input it will return plan ) < the prompt in the chain manages this ( tool ) .... Refine Code ( this can also be a tool : As it will query itself again, with the code for refinement ( the only chage is the prompt ... Tool or chain , or Graph this will always be the question ) ... So we can have a refiner prompt : the rifiner prompt could have a set of feirce requirements to folow and use for correcting code : or indsutrialising code etc :; So when we design processes etc for our agent we can also have template plans in the planner ! < the user would like to repair some code : so the planner would offer the tool choices and even disribute a boilerplate , from a set of boiler plates : ie a code planner could have many basic template to begin a project : Supplied to the model as a list of docstrings to select a template it would call the tool to return the template from file or disk storage : ( ie tools have acess to the required system and files and do not go outp=of bounds ) ( open code execution and open syystem command executions should always be human intervention as it can destruct your system : Hence Tools istead : they have boundrys : A planner could also have acces to tool collections , which can be loaded to the agent which will perform the task : ie pass a toolbox to the model ! Hence your central controller bot will access all these tools ! given a menu of tasks , which return tool collections and plans and Start nodes : allowing the model to pass a custom state to a chain and retun its output : ie create an app which does : it would ask its general planner bot : who would tell it the correct planner to begin : then that palnner would give the agent the tools and requirements and the start node ! alowing for the agent to execute the correct Graph ( tool:) with a state and expectation ! I built this process the same way as you : But after i realised the whole above discussion : and then i realized even higher perspective : that our front end should be the same as dialog flow : as this add the personality layer ( it may not be intelligent but its the constrainment we need on top) we are currently working at a very low level and should realize that we have many layers on top of this : not a UI ! between the react process (or thought process and selected methodology for the task : as the react is very good for long and indepth tasks with a graph based structiure ) but not good for general chat ! hemmce we need the front end on top of this layer , a very simple , keyword and detect/Response method : a dialog manager : as eve the rag should be plugged in at this higher level : as it is not reuired unless its for indepth tasks : as basic chat history can keep your model personalised and up to date : the rag is a short /long term memory: whihc after optimization can be finetuned into the main llm : but it still does form a part of the reseach and content process when required ( it should be a tool ( only used when required ) ! ..as a State is enough to use to perform a task and chunking and embedding for simularity should oyu be for indepth querys ! not for dowloading and sumarizing ! <the model does ot need that to sumarize data ) ( a sumarizer is only a prompt chain away : only a prompt is required ! ( so a self query tool again !) ... SO sub agents are only tools which self query the model , with clean chathistory only for the execution of the tool : only the exchange between agent and tool is saved : hence internal processing in a tool is not returned to the agent only the final output : ( it ca be shared using verbose ) ie create two reponses a verbose ( lots of lovely logging , and custom trip wire exceptions to incrase loggig potential ) or just a simple output : peserving the minimun token exchange between agents : , ie oly query , state and output... Tools ca be conditional nodes also ! so that given a input to a Tool which functions as a condotional node or routerr it will be able to have branching inside the tool ! alowing for a refiner to call a coder tool : to return the refined output to the calling agent ! ( so it branched to the coder then back to itself as the coder node always returns data to the refiner tool ! Unless specfically (boolean) told to return the output without refinement ! or we would be hiting odes whic cannot return diredct output : hence many nodes in a graph can be executed as tools: Really i hope you understand this post is an enhancement to your video and perhaps your next thought pathway: You may even find it is very less intensive on memeory as the tool unload after execution ! and your actual agent can be a very smal 1b model as the tools are the intelligence : so if you will use the saem model a moderate b is required as you will be sub calling the same model perhaps twice ( in the case of the refiner , r planner which may call a sub planner ( carfully prompted ) ) ... SO aGod 7-14b is the best model to use ! ( the newest model have been trained on function calling and tool use and planning ) !! << Concept Overview: When creating graphs or decision trees in programming, the key components are nodes and edges. Here's how they function: Nodes: Represent states, decisions, or actions. Edges: Represent the paths or transitions between these nodes. The structure typically starts with a root node and ends with a final node, forming a tree of nodes. The edges are crucial for drawing routes, checking if paths exist, or determining if they are complete. However, the actual execution of the tree doesn't depend on the edges but rather on how each node points to its subsequent node, including conditional nodes. Key Points: Execution Flow: The tree executes by starting from the root node. Each node, possibly through recursion, leads to the next, ultimately reaching the final node. Between nodes, you can pass a State-a data packet that each node processes or modifies. The final node then returns this modified state as the output. Graph Structure: Edge Map/Matrix: This is a list of lists or a dictionary that tracks all possible edges (paths) connecting nodes. For example, the edges connecting node A, B, C, etc., are mapped to allow traversal and application of algorithms like breadth-first search (BFS), depth-first search (DFS), or shortest path. For visualization, this edge list can be used to draw the graph, but the actual logic lies in how nodes are connected and executed. Graph Construction and Execution: The primary node is often an Intent Detector node, acting as a router that determines the path based on the user's intent. Each task, whether it's web research, essay writing, or coding, would have its own planner that the graph follows based on the route chosen by the intent detector. Conditional Nodes: These are special nodes that determine which path or sub-tree to follow based on specific conditions. Additional Considerations: Start Nodes: These nodes initiate the graph and can serve as entry points for different UI elements or triggers. Multiple start nodes allow for flexibility in how and where the graph begins, accommodating different types of user inputs or scenarios. Chatbot Integration: The chatbot front-end can enhance input by detecting keywords or context, thereby prompting the LLM (Language Model) appropriately. The front-end should dynamically load and unload tools based on the conversation's needs, ensuring the LLM remains responsive without being overloaded with unnecessary tools. The chatbot can also handle certain tasks directly, without needing to involve the LLM, to maintain efficiency. Final Thoughts: This approach blends traditional chatbot systems with modern LLM-based tools, creating a robust framework that combines structured dialog management with the flexibility and power of LLMs. For long-running models (e.g., 24/7 services), it's important that the graph returns to a start or reset node after completing a task, ensuring that the system is always ready for the next input. This integration of old and new methods will allow the creation of intelligent, personality-driven agents that can handle both structured tasks and dynamic problem-solving, all within a responsive and adaptive system. the point is that you dont need a langGraph or a Neo4J as well as they being the exact same thing but slightly customized specifcally for its purpose ! If you deploy jared instead as a Space in Huggingface or Runpod : then you can design a mini UI which hits your Space ! Like a chat gpt .. hence the model woul not need to tranfer so much data and you would have the speed of the space! your only processing will be in your endpoint ! so you will be using it like a perplexity and can now compare its speed and results side by side ! browser to broswer or endpoint query to query from the ui ! Sorry for the long post but i have been thinking about it indepth as you see !

  • @sai_ai-z1c
    @sai_ai-z1c 29 днів тому

    Workplaces can change when AI and analytics are combined. For this, have you looked into how SmythOS and other platforms improve AI agent collaboration? #SmythOS #Aitools

  • @dawn_of_Artificial_Intellect
    @dawn_of_Artificial_Intellect 29 днів тому

    What is your thoughts on integrating GraphRag or HybridRag?

  • @MuhanadAbulHusn
    @MuhanadAbulHusn Місяць тому

    If you don’t mind, I’m wondering about the use of LayoutPDFReader with a URL. Isn’t it designed for PDF files?

    • @Data-Centric
      @Data-Centric 24 дні тому

      It confused me initially too, it's actually a misnomer. It works for URLs too!

  • @jarad4621
    @jarad4621 Місяць тому

    You named it after me! Thanks man, it's even pronounces the same ❤

  • @nikosterizakis
    @nikosterizakis Місяць тому

    Thank you. Certainly explained quite a few things that the online documentation is missing.

  • @ashgtd
    @ashgtd Місяць тому

    Love your videos. Are you a professor?

  • @randyblasik7066
    @randyblasik7066 Місяць тому

    Cool dude. :-)

  • @SonGoku-pc7jl
    @SonGoku-pc7jl Місяць тому

    thanks! i hope this works in gpt, groq and ollama! :D thanks!

  • @peter_rabbit123
    @peter_rabbit123 Місяць тому

    Fantastic work