Thanks so much to budding AI students to share your knowledge and explain everything step by step, particularly the python functions, etc. You have a new subscriber.
Amazing research, Collab Code and overview of all the aspects. Really impactful new technology, you got a new subscription. Looking forward to more of these great videos with collab code in the future.
Would it be fairly easy to flip the RAG step round so that you prompt it with an image and the LLM is fed the associated text as part of the context? So extending the example here, a person could submit a picture of what they were wearing and it would give style tips for that?
Seems that something is returned from image search regardless of input search term. I downloaded a Kaggle data set of vehicles - trucks, cars, motorbikes and buses. I put in a search term of "Unicorn" and it still came back with images of cars. Any way to prevent this happening
Yes that’s still one of the nuances with it, a result will always be returned. What you can do is potentially a further filtering step post retrieval, either a distance cutoff or attach metadata (possible through additional kwarg with chroma) that you can filter further from as well
@@AdamLucek Thanks for that Adam, I've also put in the same question to the HELP section on their discord server. Maybe someone there can suggest something too
You could loop through a folder of photos and have a vision model provide the context for the each image as you’re vectorizing them. Then you can search your images using natural language.
We’re searching over the photos using natural language here without the need for a context generation step using zero shot image classification models, what you said is possible but an unneeded step with this method! It could be useful to do this for metadata still tho for further retrieval filtering
you are insane, because you could have just used Openclip to make the project, but you decided to train it with a clothing dataset to make it better, crazy hahaha
Thanks so much to budding AI students to share your knowledge and explain everything step by step, particularly the python functions, etc. You have a new subscriber.
Amazing research, Collab Code and overview of all the aspects. Really impactful new technology, you got a new subscription. Looking forward to more of these great videos with collab code in the future.
Great video to see expanded functionality for chroma, thx.
Fantastic demo - I have so many things to try
Great content and I learned a lot thanks and subscribed too. Keep up the good work
Would it be fairly easy to flip the RAG step round so that you prompt it with an image and the LLM is fed the associated text as part of the context?
So extending the example here, a person could submit a picture of what they were wearing and it would give style tips for that?
yes this is possible. I would look into "ActiveLoop" course on RAG where they go over a similar use-case. Good luck and have fun :)
Seems that something is returned from image search regardless of input search term. I downloaded a Kaggle data set of vehicles - trucks, cars, motorbikes and buses. I put in a search term of "Unicorn" and it still came back with images of cars. Any way to prevent this happening
Yes that’s still one of the nuances with it, a result will always be returned. What you can do is potentially a further filtering step post retrieval, either a distance cutoff or attach metadata (possible through additional kwarg with chroma) that you can filter further from as well
@@AdamLucek Thanks for that Adam, I've also put in the same question to the HELP section on their discord server. Maybe someone there can suggest something too
Great video, subscribed.
very awesome thank you for your content
Amazing! Thanks
You could loop through a folder of photos and have a vision model provide the context for the each image as you’re vectorizing them. Then you can search your images using natural language.
We’re searching over the photos using natural language here without the need for a context generation step using zero shot image classification models, what you said is possible but an unneeded step with this method! It could be useful to do this for metadata still tho for further retrieval filtering
great video!
you are insane, because you could have just used Openclip to make the project, but you decided to train it with a clothing dataset to make it better, crazy hahaha
😎🤖
♥🎯♦
Набор данных из 27 классификаций. Не проще бомжу дать ноутбук и он разметит данные за пару бутылок водки
И пропьет 500 рублёвый ноутпук