Closed NirantK closed 8 months ago
As a first step, I would implement support for both - HF and arbitrary links like it is right now. Then if we will see benefit of complete migration, we can continue
Why should we consider upgrading Huggingface Hub as a download option from JinaEmbedding specific to all Embedding models?
We'll change things in the Embedding
class. We'll add a download_from_hf
or similar function. We will continue to support existing models via GCS (arbitrary URLs via requests) in addition to Huggingface Hub.
Download Function will:
At the end of this issue, users will be able to:
This is a good reference contribution: https://huggingface.co/weakit-v/bge-base-en-v1.5-onnx/tree/main
This is what we should aim to replicate as much as we can. We'll have these models under the Qdrant Huggingface Hub account instead. So they'd be something like: qdrant/bge-base-en-v1.5-onnx
{
name: "BAAI/bge-base-en-v1.5",
sources: ["qdrant/bge-base-en-v1.5-onnx", "weakit-v/bge-base-en-v1.5-onnx"]
}
We'll have to do this for each model, one at a time:
cc @Anush008 Added the 4 models here: https://huggingface.co/Qdrant. This should unblock you completely
When a -Q and without Q suffix both are available, prefer the one without the Q suffix.
The other pattern is that of Jina: https://huggingface.co/jinaai/jina-embeddings-v2-small-en/tree/main
I'd recommend that we find a way to handle this without downloading PyTorch files at all. If we can't find that, open a new issue and I'll coordinate with the Jina folks.
is "Qdrant/bge-m3-onnx" already supported ?
The latest plans are in the most recent comment at the end
Error Handling improvements will come from two main improvements:
This issue is about the first one.
How to push models?
This is a good reference contribution: https://huggingface.co/weakit-v/bge-base-en-v1.5-onnx/tree/main
This is what we should aim to replicate as much as we can. We'll have these models under the Qdrant Huggingface Hub account instead. So they'd be something like: qdrant/bge-base-en-v1.5-onnx
We'll have to do this for each model, one at a time:
In this process, we deprecate the following models by not porting them from GCP to HF Hub on our account: