Getting Started with Ollama and Web UI

Поділитися
Вставка
  • Опубліковано 8 вер 2024

КОМЕНТАРІ • 35

  • @hfislwpa
    @hfislwpa Місяць тому +6

    2 videos in 1 day? Woah! Thanks

  • @user-zk1zm6sm2u
    @user-zk1zm6sm2u Місяць тому +2

    Interesting tutorial with Web UI and Ollama, Thanks!!!

  • @AleksandarT10
    @AleksandarT10 Місяць тому +1

    Great one Dan! Keep ups updated on the AI stuff!

  • @bause6182
    @bause6182 Місяць тому +2

    Ollama should integrate a feature like artifact that allow you to test your html css code in a mini webview

  • @user-ym6tb5xb2v
    @user-ym6tb5xb2v 25 днів тому +4

    How can I connect my local ollama3 with webUi, My webUI couldn't find the locally running ollama3

    • @MURD3R3D
      @MURD3R3D 7 днів тому

      same problem

    • @MURD3R3D
      @MURD3R3D 7 днів тому

      from home page of your webUI localhost3000 in your browser, click on your account name in the lower left, then click settings, then "models", then you can pull llama3.1 by typing it in the "pull" box and clicking the download button. when it completes, close webUI and reopen it. then i had the option to select 3.1 8B from the models list

    • @user-ym6tb5xb2v
      @user-ym6tb5xb2v 7 днів тому

      @@MURD3R3D i found that happen due to docker networking.

  • @vrynstudios
    @vrynstudios Місяць тому

    A perfect tutorial.

  • @lwjunior2
    @lwjunior2 Місяць тому

    This is great. Thank you

  • @je2587
    @je2587 22 дні тому

    Love your terminal, which tools do you use to customize it?

  • @borntobomb
    @borntobomb Місяць тому +1

    Note for 405B:
    We are releasing multiple versions of the 405B model to accommodate its large size and facilitate multiple deployment options:
    MP16 (Model Parallel 16) is the full version of BF16 weights. These weights can only be served on multiple nodes using pipelined parallel inference. At minimum it would need 2 nodes of 8 GPUs to serve.
    MP8 (Model Parallel 8) is also the full version of BF16 weights, but can be served on a single node with 8 GPUs by using dynamic FP8 (Floating Point 8) quantization. We are providing reference code for it. You can download these weights and experiment with different quantization techniques outside of what we are providing.
    FP8 (Floating Point 8) is a quantized version of the weights. These weights can be served on a single node with 8 GPUs by using the static FP quantization. We have provided reference code for it as well.
    405B model requires significant storage and computational resources, occupying approximately 750GB of disk storage space and necessitating two nodes on MP16 for inferencing.

    • @user-br4gt7xu2j
      @user-br4gt7xu2j Місяць тому

      and what about 70B? How it could be served? Could some of llama 3.1 be used by simple 16-cores laptop with integrated GPU and 32GB ram?

    • @isaac10231
      @isaac10231 5 днів тому

      When you say "we" do you work for meta?

  • @chameleon_bp
    @chameleon_bp Місяць тому +1

    Dan, what the specs for your local machine?

  • @zo7lef
    @zo7lef Місяць тому

    Would make a video on how to integrate llama 3 to wordpress website, making chatbot or co pilot

  • @trapez_yt
    @trapez_yt 26 днів тому

    Hey, could you make a video on how to edit the login page? I want to make the login page to my liking.

  • @mochammadrevaldi1790
    @mochammadrevaldi1790 7 днів тому

    in Ollama Is there an admin dashboard for tuning the model, sir?

  • @expire5050
    @expire5050 19 днів тому

    finally setup open webui thanks to you. i'd approached it, seen "docker" and left it on my todo list for weeks/months. I'm running gemma2 2b on my gtx 1060 6gb vram. any suggestions on good models for my size?

  • @NikolaiMhishi
    @NikolaiMhishi Місяць тому

    Bro you the G

  • @khalildureidy
    @khalildureidy Місяць тому +15

    Big thanks from Palestine

  • @vikas-jz3tv
    @vikas-jz3tv 15 днів тому

    How we can tune a model with custom data?

  • @DrMacabre
    @DrMacabre 25 днів тому

    hello, any idea how to set keep_alive when running the windows exe ?

  • @stoicguac9030
    @stoicguac9030 Місяць тому +1

    Is WebUI a replacement for aider?

  • @elhadjibrahimabalde1234
    @elhadjibrahimabalde1234 23 дні тому

    hello. After installing OpenWebUI, I am unable to find OLLAM under 'Select a Model'. Is this due to a specific configuration? For information, my system is running Ubuntu
    24.04.

  • @vactum0
    @vactum0 29 днів тому

    my ollama running same model is deadslow, running in laptop i5 11th gen without GPU 26GB Ram.
    Is it because of no dedicated GPU?

  • @user-km8rs4tj5w
    @user-km8rs4tj5w Місяць тому

    Thank you, I tried it but it is very slow, running it on a laptop with 16GB RAM!

  • @jaroslavsedlacek7077
    @jaroslavsedlacek7077 Місяць тому

    Is there an integration for Open WebUI + Spring AI?

  • @shuangg
    @shuangg Місяць тому +1

    6 months behind everyone else.

  • @BnmQwr-e2n
    @BnmQwr-e2n 5 днів тому

    Davis Michelle Clark Melissa Miller Frank