Closed AndyMik90 closed 3 months ago
Would be great to see how to use pinecone hybrid search with the js framework
+1 to this request!
Any updates?
You can try to implement it like this https://cookbook.openai.com/examples/question_answering_using_a_search_api
Hi, @AndyMik90,
I'm helping the langchainjs team manage their backlog and am marking this issue as stale. From what I understand, you opened this issue to request support for Pinecone hybrid search and Cohere Reranker in the JavaScript version of the repository. There has been interest and support from other users, with suggestions and examples provided for potential implementation.
Could you please confirm if this issue is still relevant to the latest version of the langchainjs repository? If it is, please let the langchainjs team know by commenting on the issue. Otherwise, feel free to close the issue yourself, or it will be automatically closed in 7 days.
Thank you for your understanding and contribution to the langchainjs project.
Hi, Cohere reranking (for Parent Retriever) is in PR - https://github.com/langchain-ai/langchainjs/pull/4738
Hi, Cohere reranking (for Parent Retriever) is in PR - #4738
@karol-f Is there any localized solution for reranking with langchain.js? I'm working on a 100% local knowledge base, so looking for a production ready and 100% local solution for reranking, in order to improve the RAG quality. Thanks.
@sugarforever Yes! There is my PR https://github.com/langchain-ai/langchainjs/pull/4738 for that. I'm already using this in my app using "Patch Package" NPM library
@sugarforever Yes! There is my PR #4738 for that. I'm already using this in my app using "Patch Package" NPM library
Cool. What's the implementation of BaseDocumentCompressor you use?
Cohere (with v2 reranking model, didn't check v3) as it is cheap and fast and gives good results.
My Rag works quite good with such setup:
For my data it works like a charm with GPT-4-turbo or Claude Sonnet. Sometimes only few, best docs are left. OFC for generating additional question I use faster and cheaper model like Haiku or GPT-3.5.
So my parent retriever chunks are:
Reranking is after:
LLM usually get 5000-15000 tokens question so it's like 1-3 cents for Claude Sonnet. In my case it's ok.
Cohere (with v2 reranking model, didn't check v3) as it is cheap and fast and gives good results. My chain - https://www.reddit.com/r/LangChain/s/xWiBObjS1z
I see. If Cohere is used, it means the data goes out to Cohere via API, if I understand correctly. Is there any alternative that can be run locally? for example open source alternatives
There are for sure os reranking models like BERT, but didn't check that
There are for sure os reranking models like BERT, but didn't check that
Thanks. I had a quick search on HF. Looks https://huggingface.co/amberoad/bert-multilingual-passage-reranking-msmarco can do the job. Any example on how to integrate it with LangChain.js?
Hi, will there be support in the js version for Pinecone hybrid search and Cohere Reranker? We got this running on Python but would love to move to Vercel AI SDK with this functionality.