Run DeepSeek R1 Privately: The Ultimate Guide

Поділитися
Вставка
  • Опубліковано 2 лют 2025

КОМЕНТАРІ • 42

  • @leonvanzyl
    @leonvanzyl  2 дні тому

    Challenge time! Share the most interesting response you got from DeepSeek R1 in the comments.

  • @gideonogunbanjo8161
    @gideonogunbanjo8161 2 дні тому

    Great content! Thank you Leon.

  • @rolandelliott9224
    @rolandelliott9224 День тому

    Thanks man, your tut was musch easier easier.

  • @francoisbeetge3881
    @francoisbeetge3881 2 дні тому

    Well done ✅ baie goed verduidelik
    befok!💪❤️

  • @user-nbfkxngjmyb
    @user-nbfkxngjmyb 2 дні тому

    very Big thanks for the great content

  • @francoisvanzyl1650
    @francoisvanzyl1650 День тому +1

    I have rtx 3080 with 10 GPU memory, i7 12700F and 32GB RAM. What parameter should I use? I know the 671B is perfect (Above ChatGPT o1 performance). I am wondering about 32b and 70b performance. How does that compare to GPT-4-turbo?

  • @burakonderuslu679
    @burakonderuslu679 День тому

    Hello, I set everything as in the video, but when i ask smth chat goes blank, it takes so long time to get responses. What can i do? I got RTX 3050, Ryzen 7 5800H

  • @rjarivi
    @rjarivi День тому

    which is best for AMD eyzen 5 3600 and 1080ti

  • @MikeNugget
    @MikeNugget День тому

    Wow, does it support file attachment?

  • @archanakamat2635
    @archanakamat2635 2 дні тому

    One more questions, Can we run 7B model in Laptops not having GPU but having internal GPU as part of processor like i3, i5, etc.? Thanks for your earlier reply. I don't see anyone as addressed my query in any UA-cam video. As i could not find it anywhere.

    • @leonvanzyl
      @leonvanzyl  2 дні тому

      Well, you need to just try for yourself.
      I was able to run the Llama 3.2: 3b model on a 10 year old laptop with an internal / onboard GPU.

    • @maury1329
      @maury1329 2 дні тому

      To do this at an "acceptable" performance, these minimum requirements might help, to start with a CPU:
      7B Params, C1, 16GB
      13B Params, C2, 32GB
      30B Params, C3, 64GB
      70B Params, C4, 128GB
      Legend
      C1: Quad-core CPU (e.g., Intel i5/Ryzen 5)
      C2: Hexa-core CPU (e.g., Intel i7/Ryzen 7)
      C3: High-end CPU (e.g., Intel i9/Ryzen 9)
      C4: Workstation-grade CPU (e.g., Xeon/Threadripper)
      Obtained via Perplexity.
      Sources checked. Please doublecheck before investing in anything.

  • @archanakamat2635
    @archanakamat2635 2 дні тому

    What is the main difference between 1.5b and 671b model and its impact on response outcomes?

    • @Swooshii-u4e
      @Swooshii-u4e 2 дні тому

      671b is the full real one with all the benefits. Anything else is just garbage and not worth it. Too many UA-cam videos of hosting small models but they don’t function like the full models

  • @archanakamat2635
    @archanakamat2635 2 дні тому

    Will the output response be different for 1.5b vs 671b also speed?

    • @Knak123
      @Knak123 2 дні тому

      Of course it will be. Check the benchmarks

  • @bassamel-ashkar4005
    @bassamel-ashkar4005 2 дні тому

    Which R1 model can I run on RTX 4090? will 32b struggle? got bad internet here and i dont want to wait and try different models

  • @Swooshii-u4e
    @Swooshii-u4e 2 дні тому

    Can I use open webui/ anythingLLM with api keys?

  • @archanakamat2635
    @archanakamat2635 2 дні тому

    Can we upload a pdf and ask questions on it in local setup?

    • @leonvanzyl
      @leonvanzyl  2 дні тому

      Yes. WebUI includes file uploads and web search.

    • @krisolak2676
      @krisolak2676 21 годину тому

      @@leonvanzyl Nope, unfortunatelly He cant read (pdf files) or describe images

  • @envoy9b9
    @envoy9b9 5 годин тому

    does it have tools? can it search the internet?

  • @Swooshii-u4e
    @Swooshii-u4e 2 дні тому

    Can you make a video of adding search + v3/r1 API keys from say hyperbolic to webui / anythingLLM (has temperature control)

  • @reidgwn7508
    @reidgwn7508 День тому

    Lekker pel

  • @The-ism-of-isms
    @The-ism-of-isms День тому

    Great video 🤗 btw I always wanna ask you flowise vs pydantic AI . I know one is no code and other is code but still wonder which framework is more powerful and worth learning. I don't know coding nor flowise but I am ready to learn any of this which is more futuristic and powerful. So plz do guide me.

    • @leonvanzyl
      @leonvanzyl  3 години тому

      Like you said, they're different as the one is no-code and the other a programming library.
      I like Pydantic, but if I had to learn one specific library, it would be Langchain and, by extension, LangGraph.

    • @The-ism-of-isms
      @The-ism-of-isms 2 години тому

      @leonvanzyl thank u for the replay , can I know why you choose those. What make it more powerful . Flowise use langchain and langraph right ? Is it better to learn both by programming or using flowise ? What u suggest

  • @justwantedtoreply
    @justwantedtoreply 7 годин тому

    nvm, didn't realise if you use the docker one you need to import a model

    • @leonvanzyl
      @leonvanzyl  3 години тому

      How else does it get access to models?

    • @justwantedtoreply
      @justwantedtoreply 2 години тому

      @leonvanzyl i tried a different one and it downloaded one by itself. I assumed this would also.

  • @Swooshii-u4e
    @Swooshii-u4e 2 дні тому

    671b is where it’s at though

  • @PugAshen
    @PugAshen 2 дні тому

    Ask it to spell strawberry 😂
    And tell it its wrong. And again. Love the internal thoughts its having and at some point just gives in

  • @pressplayg
    @pressplayg День тому

    Thank u for the video bro...is phidata better than pydantic ai ,langchain anybother frame work to build ai agent ? If have to pick one frame work to buidl ai agent which one would you pick and why 🤔🫡