Ollama - Libraries, Vision and Updates

Поділитися
Вставка
  • Опубліковано 8 чер 2024
  • Ollama Libraries: ollama.com/blog/python-javasc...
    Ollama Vision models: ollama.com/blog/vision-models
    Ollama OpenAI API: ollama.com/blog/openai-compat...
    🕵️ Interested in building LLM Agents? Fill out the form below
    Building LLM Agents Form: drp.li/dIMes
    👨‍💻Github:
    github.com/samwit/langchain-t... (updated)
    git hub.com/samwit/llm-tutorials
    ⏱️Time Stamps:
    00:00 Intro
    00:11 Ollama
    00:31 Python and JavaScript Intro
    00:56 Vision Models Intro
    01:22 OpenAI Compatibility Intro
    02:12 Demo
    05:30 Using the Python Library
    07:31 Vision Models
    10:08 OpenAI Compatibility
  • Наука та технологія

КОМЕНТАРІ • 48

  • @lucioussmoothy
    @lucioussmoothy 3 місяці тому +5

    Thanks for pulling this together. Really like the /show /save capabilities. Suggests new ways of creating and updating model files.

  • @motbus3
    @motbus3 3 місяці тому +2

    Wow. I am impressed to find one useful AI related channel.
    I mean you show things running with your code, you state real problems you find and you discuss your own results.
    Please continue with that 🙏 and thank you very much!

  • @mukkeshmckenzie7386
    @mukkeshmckenzie7386 3 місяці тому +12

    If they had an option to load multiple models at the same time (if there's enough ram/vram), it would be cool. The current workaround is to dockerize an ollama instance and run multiple of them on the same gpu.

  • @Leonid.Shamis
    @Leonid.Shamis 3 місяці тому +1

    Thank you for another very informative video. It would indeed be cool to hear more about using Ollama and local LLMs with AutoGen and for a fully local RAG system.

  • @changchoi4820
    @changchoi4820 3 місяці тому

    Wow so cool how local LLMs are progressing! So many ideas - can't handle hahah

  • @acekorneya1
    @acekorneya1 3 місяці тому +4

    would be awesome for some tutorial videos on how you make those automated tools would be awesome to know how to do things like that

  • @mr.daniish
    @mr.daniish 3 місяці тому

    the logs feature is a game changer!

  • @dllsmartphone3214
    @dllsmartphone3214 3 місяці тому +1

    ollama is the nest bro. i use it witz web ui its amazing

  • @theh1ve
    @theh1ve 3 місяці тому

    Greater canter through the recent updates, have to say I am a fan of ollama and have switched to using it almost exclusively in projects now. Not least as it's easier for others in my team to pick up. Really short learning curve to get up and running with local LLMs.

    • @samwitteveenai
      @samwitteveenai  3 місяці тому +1

      totally how I feel about it. it is simple to the point and the code is open source. I have got my team using it and everyone picks it up quickly

  • @aiexplainai2
    @aiexplainai2 3 місяці тому

    great video as always! Would you consider to cover lepton ai? looks like a great way to host llm on local machine

  • @attilavass6935
    @attilavass6935 3 місяці тому +1

    Pls. create a video about hosting an LLM server with Ollama on Google Colab (free T4) available via API. That might be a cost efficient way of hosting "local" models.

  • @ojasvisingh786
    @ojasvisingh786 3 місяці тому +1

    👏👏

  • @redbaron3555
    @redbaron3555 3 дні тому

    👏🏻👍🏻

  • @Zale370
    @Zale370 3 місяці тому +2

    Great video! Can you please cover stanfordnlp's dspy, amazing library!

    • @samwitteveenai
      @samwitteveenai  3 місяці тому

      Yeah I have been working on a few ideas for this. Anything in particular you wanted me to build etc with it?

    • @Karl-Asger
      @Karl-Asger 3 місяці тому +2

      ​@@samwitteveenaiI'll throw in a suggestion - using DSPy for an LLM agent with tool usage!
      Imo DSPy seems really powerful for bootstrapping examples for optimal answers.
      Let's say we have an LLM agent that has the purpose of performing five or six different main purposes with one or two functions for each purpose. Can use DSPY to optimize the pipeline for each of those purposes, it would be amazing.

    • @Zale370
      @Zale370 3 місяці тому

      @@samwitteveenai I'd love to see some app that uses dspy with langchain and pinecone maybe.

  • @equious8413
    @equious8413 3 місяці тому

    I serve a model with ollama and I hooked it up to a discord bot :D

  • @stephenthumb2912
    @stephenthumb2912 3 місяці тому

    Essentially this is based on llama.cpp embedded in Go but stranglely cannot handle concurrency. Love ollama and use it a lot but to run it in a production setting you have to basically spin multiple ollama server each of which can take a queue. In other words a load balancer setup with niginx or something.

  • @kenchang3456
    @kenchang3456 3 місяці тому

    I just saw on Matt Williams channel that Ollama now run on Windows natively. Just thought I'd mention it to you.

    • @samwitteveenai
      @samwitteveenai  3 місяці тому +1

      yeah I saw they have been in beta. I don't use Windows but glad it is out.

  • @supernewuser
    @supernewuser 3 місяці тому +1

    I just noticed some of these updates yesterday and it let me simplify some bits of my stack and remove litellm. It's actually kind of scary how quickly all of this stuff is becoming commodity parts.

    • @samwitteveenai
      @samwitteveenai  3 місяці тому

      totally agree things are moving so quick

  • @IronMechanic7110
    @IronMechanic7110 3 місяці тому

    Does ollama can working without internet connection when i'm using a local llm ?

    • @samwitteveenai
      @samwitteveenai  3 місяці тому

      yes it doesn't need an internet connection once you have downloaded it locally

  • @guanjwcn
    @guanjwcn 3 місяці тому

    does this mean it can run on windows now? it has been saying windows version coming soon on its website.

    • @samwitteveenai
      @samwitteveenai  3 місяці тому

      pretty sure they are still working on it and getting close.

  • @squiddymute
    @squiddymute 3 місяці тому

    can you actually stop ollama (linux) somehow ? or it runs forever and ever on the background ?

    • @notankeshverma
      @notankeshverma 3 місяці тому

      sudo systemctl stop ollama if you are using systemd.

  • @ShikharDadhich
    @ShikharDadhich 3 місяці тому

    Ollama is Awesome however there are some minor issues with oLlama:
    1. Single threaded, so can not run on server serving single url to team. It’s big issue, I don’t want everyone in my team install ollama in their machine.
    2. With Stream response its not easy to create client app as the response is not same is OpenAI
    3. CORS issue, so need a wrapping around the APIs, which means you need to install ollama and install api wrapper on every machine

  • @matikaevur6299
    @matikaevur6299 3 місяці тому

    heh, run
    ollama run llama-pro:text "what are you"
    .. about 10 times and confirm that i'm not going crazy, it's the model . . .
    that thing is outputting it's fine-tuning data verbatim .. AFAIK

  • @miladmirmoghtadaei5038
    @miladmirmoghtadaei5038 3 місяці тому

    I just don't get how it doesn't need an API for the OpenAI models.

    • @samwitteveenai
      @samwitteveenai  3 місяці тому

      its not running the OpenAI models it is using a mirror of their API to run local models

    • @miladmirmoghtadaei5038
      @miladmirmoghtadaei5038 3 місяці тому

      @@samwitteveenai thanks man. I guess I have to test it to find out.

  • @sirusThu
    @sirusThu 3 місяці тому +1

    I always thought that it is a pig

  • @Trendish_channel
    @Trendish_channel 3 місяці тому

    command line??? are you kidding?? This is super unconvinient + confusing + NOT for regular people! Not even half way close to LM Studio

    • @MarceloSevergnini
      @MarceloSevergnini 3 місяці тому

      Maybe if you actually take the time to check for yourself, you’ll notice that there is a web interface available, just need to point to your ollama instance, exactly the same as chatGPT, actually it is even better 🙃

    • @redbaron3555
      @redbaron3555 3 дні тому

      Learn CLI and stop whining.

  • @thampasaurusrex3716
    @thampasaurusrex3716 3 місяці тому

    what is better llama.cpp or ollama?

    • @Joe-yi5nv
      @Joe-yi5nv 3 місяці тому

      I'm pretty sure ollama is built on top of llama.cpp

    • @mshonle
      @mshonle 3 місяці тому

      Does Ollama support the same grammar specification that restricts your output, the way llama.cpp does? That’s a great feature which I’ve used in a project recently to force JSON output.

    • @blender_wiki
      @blender_wiki 3 місяці тому

      ​@@mshonleif you need Constrained grammars I suggest you use localAI Is very easy to implement
      local