How to Run Your Private ChatGPT-Like Assistant Offline. Run AI locally on your computer

Поділитися
Вставка
  • Опубліковано 29 вер 2024

КОМЕНТАРІ • 36

  • @Mister_MasterzRN
    @Mister_MasterzRN 13 днів тому

    What if I use the model by pasting to the Terminal, but then I no longer want that model. What if I wanted to try a different model on llama, how do I uninstall the first model that I pasted to my Terminal and replace it with a new one?

  • @AIGeek-jv1xu
    @AIGeek-jv1xu 3 місяці тому +1

    good video.
    Can we run solar 10.7B uncensored just inn the same way?

    • @bonsaiilabs
      @bonsaiilabs  3 місяці тому

      Yes, Ollama supports it. Refer to ollama.com/library/solar:10.7b. You need to pull the model first using command "ollama pull solar:10.7b" before you can select it inside openwebui. Hope this helps

  • @craigianmenzies
    @craigianmenzies 4 місяці тому +2

    Thank you very useful, if you get an error when you put in the Docker text (when using the windows version), make sure you run through all the docker install stepes and restart if needed (you will need to register).

    • @bonsaiilabs
      @bonsaiilabs  3 місяці тому

      Thank you for adding more context for people who may hit any issues with docker

  • @kierand9410
    @kierand9410 Місяць тому +1

    Can you use a voice interface wth the offline models?
    Would love to understand what's required.

    • @bonsaiilabs
      @bonsaiilabs  Місяць тому

      Yes, please look at ua-cam.com/video/RELQNYa4qNc/v-deo.html

  • @hadiaridi
    @hadiaridi Місяць тому +1

    Amazing and very clear step by step instructions! was able to replicate the work done on my computer
    Thank you so much for this excellent tutorial!

    • @bonsaiilabs
      @bonsaiilabs  Місяць тому

      You're welcome! Glad to know that it worked for you

  • @benjaminbjork842
    @benjaminbjork842 Місяць тому

    Even if I'm connected to the internet, will the model still not use the internet?

    • @bonsaiilabs
      @bonsaiilabs  Місяць тому

      Most LLMs do not use the internet. Using some OSS tools, you can create RAG based system that can fetch webpages and give the text to Local LLMs. Hope this helps

  • @priz.pov.adventures
    @priz.pov.adventures Місяць тому

    Can macbook air m1 handle this model ?

    • @bonsaiilabs
      @bonsaiilabs  Місяць тому

      Honestly, I do not know as I do not know your machine configuration. Why not try out, you will know

  • @DarkSoulGaming7
    @DarkSoulGaming7 4 місяці тому +1

    can we train this with our own personal data , and if yes how ???

    • @bonsaiilabs
      @bonsaiilabs  4 місяці тому +1

      Any open models can be fine tuned. We will make videos in future to demonstrate this use case. Thanks for asking

  • @ellied5935
    @ellied5935 4 місяці тому +1

    Thank you for this. I would like to upload/ingest files to the privategpt is that possible?

    • @bonsaiilabs
      @bonsaiilabs  4 місяці тому

      Thank you for your contribution. We just discovered privategpt, and will follow up with a video soon! Be sure to subscribe, if not already so that you get the notification when new video is live. Thank you again!

  • @mmadlang2
    @mmadlang2 4 місяці тому +1

    What's a good machine that you can recommend where I want to load llama 3 70b?

    • @bonsaiilabs
      @bonsaiilabs  4 місяці тому

      Hello, this link might be useful for your answer
      stackoverflow.com/a/78390633
      Excerpt is here
      A 70b model uses approximately 140gb of RAM (each parameter is a 2 byte floating point number). If you want to run with full precision, I think you can do it with llama.cpp and a Mac that has 192GB of unified memory, though the speed will not be that great (maybe a couple of tokens per second). If you run with 8 bit quantization, RAM requirements is dropped by half and speed is also improved.
      I hope this helps

  • @duetplay4551
    @duetplay4551 3 місяці тому +1

    hello. Can we markup the model name " llama3: latest"? thx

    • @bonsaiilabs
      @bonsaiilabs  3 місяці тому

      Hello, as per ollama.com/library/llama3, you need to use "ollama pull llama3:latest". This should work

    • @duetplay4551
      @duetplay4551 3 місяці тому

      @@bonsaiilabs thx for ur timely reply. I may ask the question in a wrong way. I should ask : can we delete this "IIama3: latest" or just leave it a blank? Because I don't wont the user know what's behind it. Thanks again.

  • @daedalusjones4228
    @daedalusjones4228 3 місяці тому +1

    Great video. Very clear. Thank you!

  • @ricardoribeiro3281
    @ricardoribeiro3281 5 місяців тому +1

    Is it possible to upload PDF files and ask for summarization?

    • @bonsaiilabs
      @bonsaiilabs  5 місяців тому +1

      Yes, it is definitely possible. Stay tuned and we will share a video about that soon

    • @bonsaiilabs
      @bonsaiilabs  5 місяців тому

      @ricardoribeiro3281, the video is almost finished and will be out in next few days. Make sure you subscribe and click the bell icon so that you get the notification once it is available. Thanks

  • @bkcamping8843
    @bkcamping8843 3 місяці тому +1

    Finally i find easy tutorial thank you

  • @autorply
    @autorply Місяць тому

    can be trained ? or will answer from his data only?

    • @bonsaiilabs
      @bonsaiilabs  Місяць тому

      You will need to fine-tune the model for your own use case. The base models can be finetuned, but you cannot train the base model as is. Hope that helps.

  • @jaspreetmaan121
    @jaspreetmaan121 5 місяців тому

    can it run without gpu

    • @bonsaiilabs
      @bonsaiilabs  5 місяців тому

      I believe it can, but the inference might be slow. I would encourage to try out and let mw know you things do with you!

    • @bonsaiilabs
      @bonsaiilabs  5 місяців тому +1

      "Ollama is popular library for running LLMs on both CPUs and GPUs". I found this reference on skypilot.readthedocs.io/en/latest/gallery/frameworks/ollama.html.
      Hope that helps!

  • @RashadPrince
    @RashadPrince 5 місяців тому

    Super cool stuff