Build an Agent with Long-Term, Personalized Memory

Поділитися
Вставка
  • Опубліковано 4 гру 2024

КОМЕНТАРІ • 54

  • @jacobgoldenart
    @jacobgoldenart 9 місяців тому +13

    Great code walk through. I'm working on something similar so it was cool to see how you approached it. Thanks for sharing.

    • @deployingai
      @deployingai  9 місяців тому +2

      That's awesome! I'd love to hear what approach ends up working for you

  • @m4tthias
    @m4tthias 9 місяців тому +6

    Nice. Look similar to Autogen's teachable agent. Appreciate your work.

    • @deployingai
      @deployingai  9 місяців тому +1

      Thanks for sharing, you're totally right. I just read through their code, it was interesting to see their prompt methods and how they implemented their memory store! Cool feature

  • @jeremybristol4374
    @jeremybristol4374 9 місяців тому

    We were just talking about this issue. All of these chat UI's treat content as if it's disposable. What works in text messaging between two people doesn't translate as well when working with computers, not if the information has utility value. This is great!

  • @ratral
    @ratral 9 місяців тому +3

    Thanks for the excellent idea and explanation.

  • @weiwei6111
    @weiwei6111 8 місяців тому +1

    Dude, your video help me a lot, THANKS!!!

    • @deployingai
      @deployingai  8 місяців тому

      That's awesome, glad to hear it!

  • @talhaanwar2911
    @talhaanwar2911 8 місяців тому +1

    your miro drawing skills are next level thing

    • @thawab85
      @thawab85 8 місяців тому +1

      i have a suspicion that @deployingai is creating AI videos to showcase his miro skills.

    • @deployingai
      @deployingai  7 місяців тому

      Haha thanks!

  • @joao.morossini
    @joao.morossini 7 місяців тому +2

    Excelent content! Thanks for sharing, man :D

  • @JulianHarris
    @JulianHarris 9 місяців тому +2

    This is so great: structured data extraction from conversations is something I’m also working on. And by the way: congrats: you have a better like-to-view ratio than MrBeast, at 4%. 🚀

    • @deployingai
      @deployingai  9 місяців тому +2

      That's cool, thanks for sharing! I'd love to hear the approach that ends up working for you.

  • @dinugakasun5218
    @dinugakasun5218 6 місяців тому +1

    Superb content, Thanks for sharing ✨

  • @zahlendatenfakten2819
    @zahlendatenfakten2819 4 місяці тому

    Great explanation, thank you!

  • @jackmartin1146
    @jackmartin1146 9 місяців тому +19

    Hello this is so cool, is it possible to share the full code including backend/front-end code as well, would love to try this, thanks so much!

  • @madelles
    @madelles 9 місяців тому

    Hey thanks for this. Very nice to see an actual application build up like this.
    I will most certainly come back for more.
    I would be interested in seeing you set up a more corporate oriented use case. What if a free text field in a form contains information relevant to a sign up for a a services or something like that

  • @databasemadness
    @databasemadness 9 місяців тому +1

    Super loaded tutorial 🎉

  • @benjaminbascary8219
    @benjaminbascary8219 7 місяців тому +2

    Can you share the vite front end? or how you setted up the front and the backend?

  • @selvakumars6487
    @selvakumars6487 4 місяці тому

    Thank you, Aiming to build a similar one for a different use case. If you can share the code with appropriate licensing would be helpful !

  • @ypindi
    @ypindi Місяць тому

    Hello Deploying AI, such an amazing tutorial and it was really useful in my project. Can you please share the code for the Supabase's Postgres database backend and the UI please?

  • @akashdeb9823
    @akashdeb9823 9 місяців тому

    I think I missed this in the memgpt paper but summarizing and storing attributes in long term memory and then refetching it for the context window is also likely going to increase the latency of the main response

    • @deployingai
      @deployingai  9 місяців тому +1

      Agreed, that's something you would need to monitor for sure. For this version, I chose to just pass the memories along in the request so they wouldn't delay the response, but that will break down once the memory store gets to a point where you need to trim it.
      I think you could always pre-fetch the entire condensed form of the conversational memory ahead of the user's request, but it might be hard to pre-fetch just the memories relevant to the new query in the small window of time between when the person finishes their message and submits their request.
      Someone pointed out AutoGen's Teachable Agent approach to memory management, and they warn: "Note that this adds extra LLM calls involving a relatively small number of tokens, which can add a few seconds to the time a user waits for each response."
      But it looks like their approach adds multiple calls before and after the main request, which are all blockers to a response.

  • @kaisaiokada776
    @kaisaiokada776 4 місяці тому

    Would you be able to share the doc of Mem GPT?

  • @tharunbhaskar6795
    @tharunbhaskar6795 9 місяців тому +7

    Interesting. Subscribed. By the way, can I get the source code for the UI. It looks pretty neat and clean

    • @deployingai
      @deployingai  9 місяців тому +1

      Great, thanks for watching! I'll need to clean it up a bit, but I'll share it when I have a chance. I used Radix UI and Tailwind for this demo, so they made this very easy to throw together.

    • @jackmartin1146
      @jackmartin1146 8 місяців тому +1

      Hello@@deployingai possible to share the code now?

    • @ypindi
      @ypindi 2 місяці тому

      Hello @deployingai - amazing tutorial. Thanks a lot! Can you please share the source code for the UI as well please it will be really useful? Thanks.

  • @outworldlyoisin
    @outworldlyoisin 2 місяці тому

    Hey, awesome work! Can you please share the UI code?

  • @khushpatel7060
    @khushpatel7060 8 місяців тому +1

    This is amazing, I creating something similar, Now I will use sentinal approach. btw, I would like ask, where can i find the frontend?

  • @PaulFischerclimbs
    @PaulFischerclimbs 9 місяців тому

    Thanks for sharing 👍

  • @paarttipaabhalaji336
    @paarttipaabhalaji336 4 місяці тому

    I have one query here. Conversation memory and context length are different ? . If the input context length of the LLM is 32k. then Prompt Input + conversation memory context length should not exceed 32k right ? please correct me if I'm wrong.

  • @unclecode
    @unclecode 7 місяців тому

    Great work, don't you mind to share the code for front end? ;)

  • @kayalvizhi8174
    @kayalvizhi8174 6 місяців тому

    Did you forget to add the graphQl frontend code to Github ?

  • @adventurelens001
    @adventurelens001 9 місяців тому

    Wow, this is super cool. Could something like this be applied to using an LLM to code a web app? One of my problems right now is that I'm using gpt4 to help code a project but before I can get through it, the context window creeps in and it starts to return incorrect code.

  • @zhrannnnn
    @zhrannnnn 8 місяців тому

    Cool channel & video!! May I ask how long it took in order to get some views on your videos? Did you got views within 24 hours? Or did you start to get some views after X uploads?

    • @deployingai
      @deployingai  7 місяців тому

      Hey thanks for watching! I've only published a handful of videos so I don't have much info for you, but the first few were definitely pretty quiet.

  • @michaelbuloichyk8986
    @michaelbuloichyk8986 8 місяців тому

    How do you store the memories ? I prabably implemented it wrong, I did a while loop with input, now trying to come up with a solution of how to actually store the messages in the memory

    • @michaelbuloichyk8986
      @michaelbuloichyk8986 8 місяців тому

      Can you share the details about how you implemented vectorDB in the loop? Would love to see the code. Much love, appreciate the effort of making the video. thanks!

    • @deployingai
      @deployingai  8 місяців тому

      In my demo, I set up a backend using Supabase's Postgres database. The agent writes the memories directly to the database, and I stream the memories from the database to the frontend to instantly see updates. And since I have the memories available already in the frontend, I just use those in my requests to the LLM so that I can avoid having to make a request to my db before every new request to the LLM.

  • @coenkuijpers1800
    @coenkuijpers1800 4 місяці тому

    Great content, not only this visit. A shame you ghosted. :(

  • @gary3548
    @gary3548 6 місяців тому

    “I won’t go through how I actually built the application” what’s the point ? I switched off here.

  • @cluelesssoldier
    @cluelesssoldier 9 місяців тому +4

    13:38 - Did... Did you just bribe and / or threaten your LLM to produce better results? lmao

    • @akashdeb9823
      @akashdeb9823 9 місяців тому +1

      He was also compassionate and asked the LLM to take a deep breath

    • @cluelesssoldier
      @cluelesssoldier 9 місяців тому

      I loved that part, as well! lol@@akashdeb9823

    • @deployingai
      @deployingai  9 місяців тому +2

      Ha! Is it weird that I feel guilty about it now?
      Thanks for pointing this out, you sent me down a rabbit hole of great papers on prompting techniques and a new idea I want to try out. If it works I'll share a video!
      I really liked this paper on eccentric prompts: arxiv.org/pdf/2402.10949.pdf

    • @cluelesssoldier
      @cluelesssoldier 9 місяців тому

      @@deployingai Can’t wait to see what else you come up with! After reviewing that document on eccentric prompts, it actually has me thinking I need to employ more ToT prompting into my own applications… 🤔

  • @squiddymute
    @squiddymute 9 місяців тому +2

    great, now do this with an open source LLM