Deploying machine learning models on Kubernetes

Поділитися
Вставка
  • Опубліковано 30 вер 2024

КОМЕНТАРІ • 51

  • @abdjanshvamdjsj
    @abdjanshvamdjsj Рік тому +9

    Brooooo this was so good.

  • @bpac90
    @bpac90 27 днів тому

    excellent!! I'm curious why my search always shows garbage and videos like this never come up. This was suggested by Gemini when I asked a question about ML model deployment.

  • @ivanxiecornell
    @ivanxiecornell Рік тому +3

    Would appreciate a video using VScode to include docker contain files, k8s file and Fast API

  • @ludwigstumpp
    @ludwigstumpp Рік тому +7

    Always a pleasure to watch someone as talented as you! Keep it up :)

  • @SunilSamson-w2l
    @SunilSamson-w2l 2 місяці тому

    the reason you got . , ? as the output for [MASK] because you didn't end your input request with a full stop. Bert Masking Models should be passed that way. "my name is [MASK]." should have been your request.

  • @kwang-jebaeg2460
    @kwang-jebaeg2460 Рік тому +1

    Look forward to show your face alot :))

  • @fizipcfx
    @fizipcfx Рік тому +2

    he is back 🎉

  • @unaibox1350
    @unaibox1350 Рік тому +1

    Amazing video. In min 5:25 how did you do to open the second bash in the console? I was searching for a long time and I can't find anything. Thanks and regards!

    • @mildlyoverfitted
      @mildlyoverfitted  Рік тому +1

      Thank you! You need to install a tool called tmux. One of its features is that you can have multiple panes on a single screen.

    • @unaibox1350
      @unaibox1350 Рік тому +1

      @@mildlyoverfitted Thank you! Will dig in it now

  • @zhijunchen1248
    @zhijunchen1248 Рік тому +1

    Hi, I would like to use GPU to accelerate this demo, can you give me some tips? Thank you

    • @mildlyoverfitted
      @mildlyoverfitted  Рік тому

      So if you wanna use minikube this seems to be the solution. minikube.sigs.k8s.io/docs/handbook/addons/nvidia/

    • @zhijunchen1248
      @zhijunchen1248 Рік тому +1

      @@mildlyoverfitted thankyou, i use the "--device" flag of transformers-cli to enable GPU. And I found that serving app takes up almost gpu memory and no compute power. Whatever, thankyou for your video!

  • @davidpratr
    @davidpratr 7 місяців тому

    really nice video. Would you see any benefit of using the deployment in a single node with M1 chip? I'd say somehow yes because an inference might not be taking all the CPU of the M1 chip, but how about scaling the model in terms of RAM? one of those models might take 4-7GB of RAM which makes up to 21GB of RAM only for 3 pods. What's you opinion on that?

    • @mildlyoverfitted
      @mildlyoverfitted  7 місяців тому +1

      Glad you liked the video! Honestly, I filmed the video on my M1 using minikube mostly because of convenience. But on real projects I have always worked with K8s clusters that had multiple nodes. So I cannot really advocate for the single node setup other than for learning purposes.

    • @davidpratr
      @davidpratr 7 місяців тому +1

      @@mildlyoverfittedgot it. So, very likely more petitions could be resolved at the same time but with a very limited scalability and probably with performance loss. By the way, what are those fancy combos with the terminal? is it tmux?

    • @mildlyoverfitted
      @mildlyoverfitted  7 місяців тому +1

      @@davidpratr interesting:) yes, it is tmux:)

  • @aditya_01
    @aditya_01 8 місяців тому +1

    great video thanks a lot really liked the explanation !!!.

  • @mmacasual-
    @mmacasual- 4 місяці тому +1

    Great example. Thanks for the information

  • @shivendrasingh9759
    @shivendrasingh9759 4 місяці тому +1

    Really helpful for foundation on ml ops

  • @JoseMiguel_____
    @JoseMiguel_____ Рік тому +1

    You're great. Thanks for sharing this in such a nice way.

  • @maksim3285
    @maksim3285 Рік тому +1

    Thank you, it helped me a lot .

  • @unaibox1350
    @unaibox1350 Рік тому

    I am having a problem in the min 18:00 the model load is being killed all the time. I tried to "minikube config set memory 4096" but still having the same problem. Any idea? I've been looking for a solution for 3 hours and there is no way

    • @mildlyoverfitted
      @mildlyoverfitted  Рік тому

      Hm, I haven't had that problem myself. However, yeh, it might be related to the lack of memory.

  • @vishalgoklani
    @vishalgoklani Рік тому +1

    Welcome back, we missed you!

  • @davidyates4857
    @davidyates4857 Рік тому +1

    Great video very informative.

  • @thinkman2137
    @thinkman2137 10 місяців тому +1

    Thank you for detail tutorial!

    • @thinkman2137
      @thinkman2137 10 місяців тому

      But torchserve now has kubernetes intergration

    • @mildlyoverfitted
      @mildlyoverfitted  9 місяців тому

      I will definitely look into it:) Thank you for pointing it out!!

  • @alivecoding4995
    @alivecoding4995 Рік тому

    What terminal application is this, with the different panels?

  • @nehetnehet8109
    @nehetnehet8109 Рік тому

    Realy goood

  • @nehetnehet8109
    @nehetnehet8109 Рік тому

    Great

  • @lauraennature
    @lauraennature Рік тому

    New video 🤩

  • @EvaKreplova
    @EvaKreplova Рік тому

    Great!

  • @kwang-jebaeg2460
    @kwang-jebaeg2460 Рік тому

    OH !!!!! Glad to meet you again !!!!

  • @evab.7980
    @evab.7980 Рік тому

    👏👏👏

  • @johanngerberding5956
    @johanngerberding5956 Рік тому

    very cool video!