Thanks, this is tremendously helpful One point to note - you need to upload the embed file, not the sentence file -> upload_file(bucket_name,embed_file_path)
Great video! What is the difference between Vertex Search service VS Vector Search for RAG application? which one is better in terms of handling better retrieval of relevant documents for RAG application where we deal with 100+ PDF documents? Can you share some insights?
Hi Janaki Ram garu Can we use for developers javascript we have provide previous code embedding store data use case generate unit test cases send to gen ai llm. Please suggest which model Rag or longchain using vector search or chroma ? Which is low cost
Thanks for the tutorial! Instead of going through the ids in the json file to fetch the sentences, is it possible to integrate those directly as metadata in the index?
Great Video. One question, I noticed you used a different model (gecko) to Gemini Pro for the embeddings. Is this ok to do? I assumed the models needed to be the same for both training and inference? Thanks again
Text embedding models are independent of LLMs. You only have to ensure that the same embedding model is used for indexing the documents and the query. This is critical to retrieving the context based on the similarity.
In your video you say "sentence_file_path". However shouldn't it be "embed_file_path" ? create_tree_ah_index function should have the GCS bucket of the embedded data and not the text with teh ids right ?
Thanks, this is tremendously helpful
One point to note - you need to upload the embed file, not the sentence file -> upload_file(bucket_name,embed_file_path)
Excelent video! Thanks for sharing the code too.
Glad it was helpful!
Very excellent Learning session Janakiram!
Great video! What is the difference between Vertex Search service VS Vector Search for RAG application? which one is better in terms of handling better retrieval of relevant documents for RAG application where we deal with 100+ PDF documents? Can you share some insights?
really helpful for understanding the concept of embedding and retrieval. Thanks.
Best tutorial. Big thanks for your shared.
Hi Janaki Ram garu
Can we use for developers javascript we have provide previous code embedding store data use case generate unit test cases send to gen ai llm. Please suggest which model Rag or longchain using vector search or chroma ?
Which is low cost
really helpful I have question i have multiple pdf files how i handel with them?
Great job! Thanks a lot. What’s the difference between this approach and using langchain?
Where can I get a copy of this notebook ?
Thanks for the tutorial!
Instead of going through the ids in the json file to fetch the sentences, is it possible to integrate those directly as metadata in the index?
Thanks for the tutorial. I am bit confused which file to be uploaded to bucket. sentence file or embedding file
Great Video, thank you soo much........
Nice. Are you ok to share the colab notebook?
Yes, sure. Please check the description. I have added the links.
Great Video. One question, I noticed you used a different model (gecko) to Gemini Pro for the embeddings. Is this ok to do? I assumed the models needed to be the same for both training and inference? Thanks again
Text embedding models are independent of LLMs. You only have to ensure that the same embedding model is used for indexing the documents and the query. This is critical to retrieving the context based on the similarity.
In your video you say "sentence_file_path". However shouldn't it be "embed_file_path" ? create_tree_ah_index function should have the GCS bucket of the embedded data and not the text with teh ids right ?
i want same thing in nodej can some one please help on which library to use
the code link u have shared is incomplete, load_file is missing and other few stuffs,
Nice Video
Excellent video - can u please do same with Langchain with retrieval
Thanks for sharing knowledge.
Can you share the notebook
Please check the description. I have added the links.
Can you please do a video on "How to use the same in Langchain with retrieval"
+1
why always python is there any way to use js?
Possible to share the notebook?
The code is available at gist.github.com/janakiramm/55d2d8ec5d14dd45c7e9127d81cdafcd and gist.github.com/janakiramm/7dd73e83c92a0de0c683ed27072cdde2
Great!