Fine-Tune Llama 3.2 Vision Model with Healthcare Images in 8 mins!

Поділитися
Вставка
  • Опубліковано 3 лют 2025

КОМЕНТАРІ • 28

  • @maxt1846
    @maxt1846 2 місяці тому

    damn legendary, mate! i was struggling with fine tuning for 3 days and your give was EXACTLY what I was looking for, keep on going!

  • @moszis
    @moszis 2 місяці тому +1

    You are an absolute legend. This is gold.

  • @tomwawer5714
    @tomwawer5714 2 місяці тому

    Thank you fantastic content!

  • @saratchai
    @saratchai 23 дні тому

    Thanks a lot!! So CooL!

  • @vacksdevlopers9234
    @vacksdevlopers9234 2 місяці тому

    Thanks !! Your videos are of great help would really appreciate if you could provide an GGUF model for the same and also a video demonstrating how to quantize a model from huggingface and run on your local ollama

  • @pioggiadifuoco7522
    @pioggiadifuoco7522 2 місяці тому +1

    Thank you Mervin for the video. Would you please tell us the cost of your fine tuning process? Thanks!

  • @softwareengineer8923
    @softwareengineer8923 2 місяці тому +1

    Such a useful content, thanks a lot for the video👍

  • @soulaimanebahi741
    @soulaimanebahi741 2 місяці тому

    thank you for the démonstration. do you think can we fine tune this model on a videos data?

  • @babusivaprakasam9846
    @babusivaprakasam9846 2 місяці тому

    Super cool.

  • @ychafiqui
    @ychafiqui Місяць тому +2

    Hello, I am facing an issue when trying to import FastVisionModel from unsloth:
    AttributeError: 'NoneType' object has no attribute 'start'

  • @kishpop
    @kishpop 2 місяці тому

    Thx mate. These are awesome. Might you be able to do a similar video for training not just the external LoRA model weights to the cloud but one for training it locally on device and saving the weights locally - potentially to be used with Ollama?

    • @kishpop
      @kishpop 2 місяці тому

      Just to add I presume the Ollama route is probably not possible with your own local version right? Or is it?

  • @tamilselvan3525
    @tamilselvan3525 Місяць тому

    How long will the whole process takes?

  • @nothingdone3011
    @nothingdone3011 2 місяці тому +1

    🎯 Key points for quick navigation:
    🦙 The video demonstrates how to fine-tune the Llama 3.2 Vision model, a powerful 11 billion parameter AI specifically designed for healthcare image analysis.
    📸 The aim is to improve the model's accuracy and specificity in diagnosing from medical images like radiographs by fine-tuning it with specialized data.
    🔧 The fine-tuning process is made simpler and faster with the use of the "unso" tool, enabling efficient model adjustments.
    🩻 A specific medical dataset, "Radiology mini," containing images and captions, is used to train the model to provide accurate diagnostic descriptions.
    💾 The workflow involves several steps: loading the model, dataset preparation, pre-training evaluation, model training, and post-training assessment.
    🖥️ The video explains the technical setup required, including software installation, model importing, and use of specific configurations.
    ⏳ During training, the model's performance is monitored, and after training, the model's accuracy in image description improves notably.
    ☁️ Once fine-tuning is complete, the adjusted model is saved and uploaded to Hugging Face, integrating changes for wider access and application.
    Made with HARPA AI

  • @philtoa334
    @philtoa334 2 місяці тому

    Nice.

  • @alexa1017
    @alexa1017 2 місяці тому +2

    I am wondering about the ToS of Llama for healthcare.. (clearly, it is non-commercial use…)

    • @rousabout7578
      @rousabout7578 2 місяці тому

      Top result on Google search. "If, on the Llama 3.2 version release date, the monthly active users of the products or services made available by or for Licensee, or Licensee’s affiliates, is greater than 700 million monthly active users in the preceding calendar month, you must request a license from Meta, which Meta may grant to you in its sole discretion, and you are not authorized to exercise any of the rights under this Agreement unless or until Meta otherwise expressly grants you such rights."

    • @mik3lang3lo
      @mik3lang3lo 2 місяці тому

      I was thinking the same thing

    • @rousabout7578
      @rousabout7578 2 місяці тому

      Google 'Llama 3.2 licence'. It's fairly open.

  • @yotubecreators47
    @yotubecreators47 2 місяці тому +3

    Any time I see Fine tuning I click like before watching the video we need more finetuning + CPT cont. pretraining etc..
    how to epxlain loss/validation charts & W/B

  • @Shine-and-line
    @Shine-and-line 2 місяці тому

    Very useful information, thanks. Can you please share the code?

    • @MervinPraison
      @MervinPraison  2 місяці тому

      Sure: mer.vin/2024/11/finetune-llama-3-2-vision-radiology-images/

  • @moszis
    @moszis 2 місяці тому

    Do you know of a way to run a fine tunned vision model locally? Looks like saving as gguf is not supported for vision models, so cant run it on ollama.

    • @Aman-kb9gb
      @Aman-kb9gb Місяць тому

      You can make a symbolic link between your local llama model and ollama. After that you will be able to run

  • @robertjalanda
    @robertjalanda 2 місяці тому

    Thanks for this video. Is there an example or dataset with a more broad usecase? Radiography is a bit niche

  • @SheldonCooper0501
    @SheldonCooper0501 Місяць тому

    Anyone facing issue running unsloth FastvisionModel in Mac?