Closed fsiefken closed 1 week ago
Absolutely! I was actually investigating it yesterday to set it up.
Embedding wise LanceDB only supports OpenAI.... while Qdrant appears to support everything. https://lancedb.github.io/lancedb/embeddings/embedding_functions/ https://qdrant.tech/documentation/embeddings/
I wonder could Ollama provide embeddings to LanceDB through an OpenAI compatible api? https://ollama.com/blog/embedding-models https://github.com/ollama/ollama/issues/2416
It looks like you can define a schema and store the vectors. I think we outperform the embedding speed of lancedb if we do it ourselves 👯
While Qdrant has local support and provides a docker image, LanceDB, another Rust vector database, runs serverless and in process, similar to Sqlite. Qdrant has no GPU support yet when building the index, LanceDB does. https://lancedb.github.io/lancedb/ann_indexes/#creating-an-ivf_pq-index
Query wise LanceDB appears to be as fast or faster then other vector databases and has lower memory requirements due to it being serverless. https://blog.lancedb.com/benchmarking-lancedb-92b01032874a/ https://github.com/prrao87/lancedb-study
These can potentially speed up RAG pipelines compared to using Qdrant https://blog.lancedb.com/accelerating-deep-learning-workflows-with-lance/
Rust API pointers: https://docs.rs/lancedb/latest/lancedb/ https://lancedb.github.io/lancedb/reranking/ https://towardsdatascience.com/scale-up-your-rag-a-rust-powered-indexing-pipeline-with-lancedb-and-candle-cc681c6162e8