Open mballav opened 1 month ago
Hey there, @mballav! 🎉 I'm Dosu, your friendly neighborhood bot, here to help you tackle bugs, answer questions, and guide you on your journey to becoming a super contributor while we wait for a human maintainer to chime in. Excited to dive into your issue about using Milvus' built-in embedding models with LlamaIndex. Give me a sec to gather some info and I'll get back to you with what I find!
You can't use milvus's built-in embeddings, but you can
embed_model = OpenAIEmbedding(..., embed_batch_size=2000)
@logan-markewich, Thanks a lot for the prompt reply.
Looking at the options that you mentioned, option #3 is appealing, though even zilliz managed index requires openai key. I am assuming that zilliz under the hood vectorizes a document using OpenAI's text-ada model. If we go that route, do you think that their pipeline would be any faster than if we ourselves vectorized using OpenAI?
The Zilliz managed index pipeline, designed for use with LlamaIndex, does indeed necessitate an OpenAI key, indicating its reliance on OpenAI's services for tasks such as generating embeddings. While it's not explicitly confirmed that it uses OpenAI's text-ada model for vectorization, the managed nature of Zilliz's service suggests potential optimizations and features that could lead to faster document vectorization compared to direct usage of OpenAI. These include managed service integration, pipeline optimization, and automatic management features that are tailored for efficiency.
To accurately assess whether Zilliz's pipeline offers a speed advantage over direct OpenAI vectorization, it would be beneficial to conduct performance benchmarking specific to your use case. The LlamaIndex repository provides tools for such benchmarking, which could serve as a basis for your tests.
For detailed insights and potential performance benchmarks, you might explore the following resources within the LlamaIndex repository:
This approach will help you determine the potential speed improvements of using the Zilliz managed index pipeline over a direct, self-managed implementation with OpenAI.
To continue the conversation, mention @dosu.
@logan-markewich, Thanks a lot for the prompt reply.
Looking at the options that you mentioned, option #3 is appealing, though even zilliz managed index requires openai key. I am assuming that zilliz under the hood vectorizes a document using OpenAI's text-ada model. If we go that route, do you think that their pipeline would be any faster than if we ourselves vectorized using OpenAI?
@mballav zilliz managed index can use a hosted version of bge models, checkout this for all supported models.
Question Validation
Question
We are considering using Milvus on the cloud (zilliz.com) as a managed solution for storing and querying our vectors. Right now, if we use Open AI to get our documents vectorized, it takes a long time if there are more than say 10 documents.
Also, we are using llama-index (over langchain) for our RAG application.
All the examples I found, that showcase use of Milvus vector store, use Open AI for embedding. Is there a way to use Milvus'es built-in local models for embedding which will likely speed up the document uploads.
Any pointers or any alternative solution will be greatly appreciated.