GPU vs CPU: Running Small Language Models with Ollama & C#

Поділитися
Вставка
  • Опубліковано 5 лют 2025
  • In this video, we'll explore the performance differences when running Large Language Models (LLMs) in Ollama using both the CPU and GPU. Watch as I demonstrate a live sample in C# using Microsoft.Extensions.AI to run Ollama inside a Docker container. Curious to see how these models perform locally? Let's dive in and compare the results!
    Useful links:
    .NET Video Analyzer repository: aka.ms/netaivi...
    Ollama in Docker: ollama.com/blo...
    .NET & AI Show: • .NET AI Community Stan...

КОМЕНТАРІ • 15

  • @bdanuw
    @bdanuw 3 місяці тому +1

    Another great video from ElBruno :)
    Thank you Bruno.

    • @elbruno
      @elbruno  3 місяці тому

      Glad you liked it!

  • @eugene5096
    @eugene5096 3 місяці тому +1

    Thank you Bruno, interesting as usual !!!

    • @elbruno
      @elbruno  3 місяці тому

      @@eugene5096 Thanks! The CPU vs GPU is a wow one 😁

  • @RomuloMagalhaesAutoTOPO
    @RomuloMagalhaesAutoTOPO Місяць тому +1

    👍

  • @muluhsenet7582
    @muluhsenet7582 2 місяці тому +1

    ❤❤❤

  • @bilalbilal7674
    @bilalbilal7674 3 місяці тому +1

    Bilal here😊, I think you should go for creating an extension then it will be good and easy to accessible

    • @elbruno
      @elbruno  3 місяці тому

      There is one in the Aspire Community Toolkit: github.com/CommunityToolkit/Aspire/tree/main
      I may record a video about that one!
      Best

  • @samirou976
    @samirou976 3 місяці тому

    Hi ! This is very interesting but I wonder how it would perform on an NPU ? Is it possible to make it run on NPU ?

    • @cuachristine
      @cuachristine 3 місяці тому +1

      NPU is a GPU with all the graphics bits removed.

    • @samirou976
      @samirou976 3 місяці тому

      @@cuachristine Yes I know thank you but that was not my question

    • @elbruno
      @elbruno  3 місяці тому +1

      Ohh that's a great question! I still don't have access to an NPU machine, however, if docker desktop allows access to the NPU, it should work. Let me ask Justin (a fellow CA.NET), who rocks the docker world to see what he can share about this.

    • @samirou976
      @samirou976 3 місяці тому

      @@elbruno Thanks man I would realy apreciate that 🙂

  • @jimmymac601
    @jimmymac601 3 місяці тому

    Does this support multiple GPUs ?

    • @elbruno
      @elbruno  3 місяці тому

      I'm not sure, I'll say no, we may need to check with the ollama team