BabyAGI: Discover the Power of Task-Driven Autonomous Agents!

Поділитися
Вставка
  • Опубліковано 9 кві 2023
  • Colab: colab.research.google.com/dri...
    Paper: yoheinakajima.com/task-driven...
    Github: github.com/yoheinakajima/babyagi
    :robot_face:
    In this video, I look at BabyAGI a demo application from the paper - Task-driven Autonomous Agent Utilizing GPT-4, Pinecone, and LangChain for Diverse Applications.
    My Links:
    Twitter - / sam_witteveen
    Linkedin - / samwitteveen
    Github:
    github.com/samwit/langchain-t...
    github.com/samwit/llm-tutorials
  • Наука та технологія

КОМЕНТАРІ • 59

  • @jameshughes3014
    @jameshughes3014 Рік тому +14

    This channel is so refreshing. In a world filled with hype, you give us just the facts and a very clear overview and I'm very grateful.
    I can't wait to see one of these auto-language model thingies use a local offline LLM to do something useful.
    I guess I gotta return my candles now that I know I can't bring them in with me.

  • @narutocole
    @narutocole Рік тому +8

    Dude I love how you always include a colab notebook to experiment with!

    • @samwitteveenai
      @samwitteveenai  Рік тому +3

      Thanks Jordan, I am big believer that you really only know how something works with code, Colab is great for quick ideas and trying things out etc. and these things can be pain to set up envs etc. so hopefully this way people and check it out quickly themselves.

    • @unathimatu
      @unathimatu Рік тому +2

      @@samwitteveenai This is how research works and we love it!!!!!!

    • @srikarpamidi1946
      @srikarpamidi1946 Рік тому

      Seriously...its awesome...

  • @joewardpr
    @joewardpr Рік тому +1

    Thank you, Sam. That was very helpful to explain what BabyAGI is all about. Essentially, it's a task list builder with iterative refinement. When the algorithms are improved, there's definitely some value to be had here. I think we'll end up seeing the autonomy capped and more "task structure patterns" introduced. It's a great initial proof of concept that has been spawning others to try their hand with differing approaches. TBD.

  • @sitedev
    @sitedev Рік тому +8

    This is nuts. In my mind I visualize these agents as complex text filters with fuzzy logic - at least that's how I explain it to my wife during my endless (positive) rants about AI. Still married after 30 years but the endless rants are probably stretching the relationship a little more than usual. I feel like a boy that just received a box of magical lego bricks.

  • @MadhavanSureshRobos
    @MadhavanSureshRobos Рік тому

    So glad subscribed to you! Wonderful content!

  • @riggitywrckd4325
    @riggitywrckd4325 Рік тому +4

    Someone is working on getting this going with your local textgen instance. Who ever wants to work on this should. It will be powerful to be able to run something like this on your home computer. Llama is pretty smart and should give some very good results. Lets hope it comes out soon. If someone is in the process of getting this to work locally I would love to help in what ever way I can :)

    • @samwitteveenai
      @samwitteveenai  Рік тому +1

      You could do this pretty nicely with Vicuna or Koala and then make it even stronger with finetuning them.

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w Рік тому +24

    Sometimes it’s hard to know what the differences are given AutoGPT, Jarvis, Langchain, BabyGPT etc.

    • @witness1013
      @witness1013 Рік тому

      No, it's not

    • @lostwave1748
      @lostwave1748 Рік тому +9

      @@witness1013 OK

    • @lordanubi
      @lordanubi Рік тому +9

      From my understanding:
      AutoGPT - chain of chatgpt thoughts and execution of code/commands
      Baby AGI - chain of chatgpt tasks executed in parallel for faster output
      Jarvis/HugginGPT- chain of multiple hugging face models (not only gpt also image, body recognition and so on) to form a response to initial prompt
      The first two are both based on LangChain, which is a tool to make AI interacts with itself.
      Not sure about Jarvis but it might be since ChatGPT plug-ins are based on that and Microsoft people is playing a lot with it
      Guess next step is just merging them all together, chain of multiple different models together (in parallel or one after each other) which can also perform direct actions in the computer

    • @lordanubi
      @lordanubi Рік тому +5

      Oh and one important thing about babyagi is that it uses pinecone to remember the first part of the tasks you assigned to prevent token limits. This is not done in autogpt I think.
      What we’re also missing for final agi is MemoryGPT, which can make you decide if you want ai to persistently remember information about you or your computer, and also a nice web interface. Web UI for text generation is a good starting point

    • @Pabz2030
      @Pabz2030 Рік тому +1

      @@lordanubi Auto GPT now uses Pinecone

  • @FrancoisPesce
    @FrancoisPesce Рік тому +5

    I've spent a day playing on this code Saturday after playing with Auto-GPT.
    Few comments: It had a few bugs when I tested it. Some tasks generated more tasklist, and where not identified as such, and they are later caught up by their prioritizer creating a bizarre pattern of tasks having several ids and numbers in the begining of their name.
    Also, to avoid loop, there's easy optimization to just locally store a cache that keep result of past tasks, and preventing them to be re-inserted.
    I'm not a big fan of adding yet another API dependency (the pinecone one), as far as I understand, they're just using it to do concept search using OpenAI embeddings (i.e., they vectorized result of a task and store in pinecone). Unless you're planning to have a huge task like millions of documents, there's no need of an elastic resource here, you might as well run a local vector storage. After a few hours, I ended up implementing a containerized local weaviate to replace pinecone in their code. I'm curious why Auto-GPT and this project keep throwing pinecone in their dependencies...

    • @samwitteveenai
      @samwitteveenai  Рік тому +3

      These are all good points. I didn't want to beat up on the project, but I also felt a bit underwhelmed and that this could be done easier in many ways. Agree about the Pinecone stuff - I like pinecone as a service, but I I am not sure it is really needed here.

    • @terbospeed
      @terbospeed Рік тому +1

      Yea I think chromadb would also work well. Pinecone certainly has a lot of coverage as the go to for this, and while it is pretty simple to setup, it is curious how quickly everyone is to hand over all of their data to everyone :(

    • @memesofproduction27
      @memesofproduction27 Рік тому

      @@samwitteveenai First of all, excellent content, thank you for putting this out there. What's your opinion on the data privacy w.r.t using pinecone as a vector store, given that they host and control everything?

    • @samwitteveenai
      @samwitteveenai  Рік тому

      Lately I have been hearing they are losing people's data, so I would say make sure you have a backup. twitter.com/daraladje/status/1631050397841342465?t=41FRkMxs57jhqvKUa_JPqQ&s=19 You can also run open source versions which are probably just as good.

  • @Null-Red-Blue
    @Null-Red-Blue Рік тому

    Babyagi can be done in ChatGPT with plugin access. Auto-GPT actually performs actions instead of basic web scraping with tasks.

  • @DavBotsArcade
    @DavBotsArcade 11 місяців тому

    what I find interesting in these applications is "hallucinations" AutoGPT for example, commands have defined arguments, but quite often the model needs prompt reinforcement or it quite often uses incorrect arguments. I wonder if some of these projects overkill the use of the LLM and some functions like sorting or specific arguments to a command could be handled by "dumb" functions.

  • @gridplan
    @gridplan Рік тому +1

    How would one go about interacting with this programmatically, say from a microservice?

  • @smartapp9534
    @smartapp9534 11 місяців тому

    any chance to produce similar videos for SuperIGA? a video like this can greatly influence developer's preferences or even decision amoung similar agent sytems such as gptagent vs superiga vs this one (babyiga), due to lack of docs and examples; this channel has explained it so well that BabyIGA becomes my preference as I had many failure attempts with SuperIGA, even though I'm more attracted by the latter's concepts,I couldn't find good info about it, while I lack the ability & time to figure it out.

  • @KA-kp1me
    @KA-kp1me Рік тому +1

    It will only be useful when people will have full ability to run stuff like that locally. Otherwise sharing personal details with it, is a bad bad idea. @Sam Witteveen, have you managed to run it with LLAMA?

  • @mysticaltech
    @mysticaltech Рік тому

    Yeah, I wouldn't give it my CC too, as it would max it out in one go! 🤣

  • @Haz2288
    @Haz2288 Рік тому +1

    I tried to demo this for my wife and of course pinecone was down 😩

  • @henkhbit5748
    @henkhbit5748 Рік тому

    How much tokens it used in the end and thus $$? I think openai$$$ loves this kind of initiatives because they will generate extra money with their verbose answers....

    • @samwitteveenai
      @samwitteveenai  Рік тому

      yeah if you made it to the end you will have notice i aborted it as I felt it was just burning tokens etc.

  • @joannot6706
    @joannot6706 Рік тому +2

    Do people realise that this is basically the implementation of the theory of the bicameral mind from the 70's?
    Featured in the 2016 TV show westworld season 1 were robots reach consciousness?
    It's a theory were one part of the brain which appears to be speaking giving orders, and a second part which listens and obeys.
    Isn't it funny? AutoGPT also kind of works the same.

    • @samwitteveenai
      @samwitteveenai  Рік тому +2

      If you think this is Westworld wait till you see "Generative Agents: Interactive Simulacra of Human Behavior" it even has the reveries!!

  • @terbospeed
    @terbospeed Рік тому

    I've been running babyagi over API to the open assistant 13b model, and it does pretty well, I am interested in melding the profession descriptions from CAMEL, or maybe just modifying CAMEL to support offline processing.

  • @timetravellingtoad
    @timetravellingtoad Рік тому +2

    Thanks! How does this stack up to AutoGPT? Also, is there a way to use it on a local LLM?

    • @samwitteveenai
      @samwitteveenai  Рік тому +6

      I am doing AutoGPT probably next up and I want to do a video comparing all these. They all have interesting bits but are also over hyped I feel.

    • @XHackManiacX
      @XHackManiacX Рік тому +1

      @@samwitteveenai They are very over-hyped atm. I've seen lots of "BabyAGI/AutoGPT have sparks of AGI" videos lately.
      Probably just because people don't understand what they're doing under the hood.
      It would be cool if you made a video explaining what they're actually doing in layman's terms, so maybe then the average person can understand that it's not magic lol
      Even this kind of video is too advanced for the average Joe.

    • @fontende
      @fontende Рік тому +4

      Like an OpenAi surrogates, again, even quite far away from Llama, Alpaca and etc. Local here is only your wallet, you're keep using OpenAi services but with Microsoft search engine with full identification, location and etc.
      Sabrina from Answer in progress channel showed in her videos a years ago how collecting data with Ai bot got her blocked at Google, with forever captcha linked to your ip address. She collecting a simple stat data for videos.
      It's getting closer to the serious topic of who owns data and the open source community will be forced to make very sophisticated techniques to get access to what was collected or stolen by search robots of huge corporations.

    • @samwitteveenai
      @samwitteveenai  Рік тому +1

      @Hack Slash good feedback. I have deliberately stayed away from terms like 'have sparks of AGI' at the cost of getting views because I think these things are getting over hyped.

  • @JTMoustache
    @JTMoustache Рік тому +1

    "I dont think this is approaching agi in any way" - see you in a few month

    • @joewardpr
      @joewardpr Рік тому

      I'm not sure it's actually attempting to be an AGI (other than recognizing the value of the label right now for marketing). However, it might end up being the core pattern that heads in that direction anyway. Human thought is a constant lookup of related ideas-but, instead of biasing the selection based on our experience (weighted information, and weighted information connections), an LLM more or less has a single, repeatable response to every input, and it won't change until the model is updated/refreshed. Humans change dynamically as our sensory organs are like dynamically updating weights/parameters, and we're heavily biased upon recent information. In a way, we have built in randomization. ;)

  • @MadhavanSureshRobos
    @MadhavanSureshRobos Рік тому

    Next up, we rewrite AGI to work on Vicuna and make explosive noises in the world! 🤣

  • @Wlodzislaw
    @Wlodzislaw Рік тому

    Would you expect more from a baby?

  • @dr.mikeybee
    @dr.mikeybee Рік тому +1

    I'm not sure Pinecone is needed for this. A Python list would work just as well for semantic search.

    • @fontende
      @fontende Рік тому

      Good luck with requesting Google servers outside browsers to get your ip in global block lists.

  • @incrediblyStupid678
    @incrediblyStupid678 Рік тому +1

    Uhm, this is pretty much exactly what auto-gpt does which was released like two weeks ago or so? Including the usage of Pinecone, etc.

  • @ranu9376
    @ranu9376 Рік тому

    First View!

  • @microgamawave
    @microgamawave Рік тому

    And I'm the second one

  • @jaysonp9426
    @jaysonp9426 Рік тому +1

    I wish people wouldn't say things like "I don't think this a approaching AGI in anyway" without defining it first. By my definition we're already at AGI. It's artificial...it uses tools without instruction and has a feedback loop that let's it improve on itself. I've yet to have anyone tell me how humans are substantially different other than being organic and having an insanely efficient processing system. If you want to say that your idea of AGI is conscious then you need to state that as part of your definition. Otherwise, it just sounds like humans being arrogant about what we are. We're not special and that fact is going to very present soon.

  • @avi7278
    @avi7278 Рік тому

    These "AGI"s are over hyped garbage. Anyone who more than five minutes with them sees the novelty wear off. They can't do anything even slightly complex and most of what they actually can do is done more easily, and faster with just straight up chat GPT with the slight advantage that they can (when they don't fail) scape the web for info. But again with chat GPT plugins the same will soon be available to everyone and bing and phind already currently hook GPT up to the internet. It's interesting to see a LLM reasoning through some basic tasks but nothing more.

    • @fontende
      @fontende Рік тому

      You don't understand, it's same ChatGPT but with that plugins you are taking about, you are using openai identification key for any work of it in Internet through Microsoft search servers, because Google banning machine search requests by command line. It's impossible to use them outside big brother.

    • @avi7278
      @avi7278 Рік тому +1

      @@fontende what don't i understand? I think it's more likely you didn't understand my comment.