wow this is too easy to be real. i am using opencv to record videos of flying saucers. i could record images and use llama to verify if there is a flying saucer in it. can i also search videos with videos: instead of images:?
@@sumukhas5418 Thanks for the answer :) Actually I am trying pytesseract to read id-card information, which are photographed by a phone and the results are not very good :/ Do you have some ideas, how I could get some better results?
Help me out ,it took less than 10 seconds to get the output , but for me it is like taking 3mins to run , of course it runs , i am happy but it is too late
Is ollama and llava is free to use and I have spec 16GB/1TB RTX 3050Ti what no. of model is suitable for my device 13B one or else. And I already using ollama basic 4GB model in my device is it ok to run 13B model and some Other model like OpenAi or Gemini API??
With 4-bit quantization, for LLaVA-1.5-7B, it uses less than 8GB VRAM on a single GPU, typically the 7B model can run with a GPU with less than 24GB memory, and the 13B model requires ~32 GB memory. You can use multiple 24-GB GPUs to run 13B model
Awesome, man! I was not aware of customizing Ollama with this kind of Python script! Thanks :)
This is quite useful!
It gives me some great ideas for my own local apps!
wow this is too easy to be real. i am using opencv to record videos of flying saucers. i could record images and use llama to verify if there is a flying saucer in it. can i also search videos with videos: instead of images:?
Thanks :) Is it possible to use this model as an ocr alternativ to get for example informationen from a jpeg image which is an id-card ?
This will be too much heavy for just that
Instead considering yolo would be a better option
@@sumukhas5418 Thanks for the answer :) Actually I am trying pytesseract to read id-card information, which are photographed by a phone and the results are not very good :/ Do you have some ideas, how I could get some better results?
Thanks for your help you legend
rag - webcam - selfawareness - speech --> tutorial pls
Help me out ,it took less than 10 seconds to get the output , but for me it is like taking 3mins to run , of course it runs , i am happy but it is too late
My computer takes more than an hour , the system is installed with a 4GB 3060 GPU , what can I do
@@santhosh-j7e I dont know man , i was like working it for my hackathon , i tried like all pc ,like pentium , i3 , i5 ,i7 but no difference.
7.5 Gb ?????
It's 4.7gb for 7b version
This was very helpful, my first time getting results from a multimodal LLM directly using Python.
If my local ram is 8 gb, which ollama model would you recommend to use?
deepseek-coder ❤
deepseek-coder ❤
Is ollama and llava is free to use and I have spec 16GB/1TB RTX 3050Ti what no. of model is suitable for my device 13B one or else. And I already using ollama basic 4GB model in my device is it ok to run 13B model and some Other model like OpenAi or Gemini API??
can we get the answer in different languages as per the client requrement just like in hindi or tamil or japanese etc if possible
is this fully offline? I am not sure you downloaded the 13B 7.4Gb package
What a nice vid. Can I do a ai without using open ai ?
Nice, very helpful!
Is it possible to create embeddings of pictures with the model?
Are there models that recognize a photo and then vectorizes it?
Thanks for the video, how to make sure that I install Ollama on the GPU not on the CPU?
how to add long term memory in this local llm ???
Riding the awesomeness wave again!
oh im too fast
How much RAM and VRAM needed ?!
With 4-bit quantization, for LLaVA-1.5-7B, it uses less than 8GB VRAM on a single GPU, typically the 7B model can run with a GPU with less than 24GB memory, and the 13B model requires ~32 GB memory. You can use multiple 24-GB GPUs to run 13B model
First comment 😊😊😊