Image Recognition with LLaVa in Python

Поділитися
Вставка
  • Опубліковано 28 вер 2024

КОМЕНТАРІ • 30

  • @yuvrajkukreja9727
    @yuvrajkukreja9727 Місяць тому +1

    Awesome, man! I was not aware of customizing Ollama with this kind of Python script! Thanks :)

  • @blackstonesoftware7074
    @blackstonesoftware7074 2 місяці тому

    This is quite useful!
    It gives me some great ideas for my own local apps!

  • @fastmamajama
    @fastmamajama Місяць тому +1

    wow this is too easy to be real. i am using opencv to record videos of flying saucers. i could record images and use llama to verify if there is a flying saucer in it. can i also search videos with videos: instead of images:?

  • @wasgeht2409
    @wasgeht2409 3 місяці тому +1

    Thanks :) Is it possible to use this model as an ocr alternativ to get for example informationen from a jpeg image which is an id-card ?

    • @sumukhas5418
      @sumukhas5418 3 місяці тому

      This will be too much heavy for just that
      Instead considering yolo would be a better option

    • @wasgeht2409
      @wasgeht2409 3 місяці тому

      @@sumukhas5418 Thanks for the answer :) Actually I am trying pytesseract to read id-card information, which are photographed by a phone and the results are not very good :/ Do you have some ideas, how I could get some better results?

  • @aaronbornmann9835
    @aaronbornmann9835 9 днів тому

    Thanks for your help you legend

  • @antonpictures
    @antonpictures 2 місяці тому

    rag - webcam - selfawareness - speech --> tutorial pls

  • @naturexmusic2567
    @naturexmusic2567 29 днів тому

    Help me out ,it took less than 10 seconds to get the output , but for me it is like taking 3mins to run , of course it runs , i am happy but it is too late

    • @santhosh-j7e
      @santhosh-j7e 8 днів тому +1

      My computer takes more than an hour , the system is installed with a 4GB 3060 GPU , what can I do

    • @naturexmusic2567
      @naturexmusic2567 8 днів тому

      @@santhosh-j7e I dont know man , i was like working it for my hackathon , i tried like all pc ,like pentium , i3 , i5 ,i7 but no difference.

  • @Justwil07
    @Justwil07 Місяць тому

    7.5 Gb ?????

  • @jaykrown
    @jaykrown 10 днів тому

    This was very helpful, my first time getting results from a multimodal LLM directly using Python.

  • @giovannicordova4803
    @giovannicordova4803 3 місяці тому +1

    If my local ram is 8 gb, which ollama model would you recommend to use?

  • @brpatil_007
    @brpatil_007 Місяць тому

    Is ollama and llava is free to use and I have spec 16GB/1TB RTX 3050Ti what no. of model is suitable for my device 13B one or else. And I already using ollama basic 4GB model in my device is it ok to run 13B model and some Other model like OpenAi or Gemini API??

  • @rajm5349
    @rajm5349 Місяць тому

    can we get the answer in different languages as per the client requrement just like in hindi or tamil or japanese etc if possible

  • @potatoes1000
    @potatoes1000 2 місяці тому

    is this fully offline? I am not sure you downloaded the 13B 7.4Gb package

  • @Isusgsue
    @Isusgsue 3 місяці тому

    What a nice vid. Can I do a ai without using open ai ?

  • @AlissonSantos-qw6db
    @AlissonSantos-qw6db 3 місяці тому

    Nice, very helpful!
    Is it possible to create embeddings of pictures with the model?

  • @derekchance8197
    @derekchance8197 Місяць тому

    Are there models that recognize a photo and then vectorizes it?

  • @R8R809
    @R8R809 3 місяці тому

    Thanks for the video, how to make sure that I install Ollama on the GPU not on the CPU?

  • @yuvrajkukreja9727
    @yuvrajkukreja9727 Місяць тому

    how to add long term memory in this local llm ???

  • @GuillermoGarcia75
    @GuillermoGarcia75 3 місяці тому

    Riding the awesomeness wave again!

  • @arjuntt2604
    @arjuntt2604 3 місяці тому

    oh im too fast

  • @aoa1015
    @aoa1015 3 місяці тому

    How much RAM and VRAM needed ?!

    • @RedFoxRicky
      @RedFoxRicky 3 місяці тому

      With 4-bit quantization, for LLaVA-1.5-7B, it uses less than 8GB VRAM on a single GPU, typically the 7B model can run with a GPU with less than 24GB memory, and the 13B model requires ~32 GB memory. You can use multiple 24-GB GPUs to run 13B model

  • @syedmokarromhossain4867
    @syedmokarromhossain4867 3 місяці тому

    First comment 😊😊😊