Open DavidPeleg6 opened 3 months ago
@DavidPeleg6 - do you have capacity to add this feature? I can help guide you. It should not be too hard
@DavidPeleg6 Here is my quick sketch at adding support for Qwen2 embedding models https://github.com/vllm-project/vllm/pull/5611
This is not sufficient though since that model you shared still says it is for causal language modeling, rather than just Qwen2Model
@DavidPeleg6 Here is my quick sketch at adding support for Qwen2 embedding models #5611 这是我添加对 Qwen2 嵌入模型的支持的快速草图 #5611
This is not sufficient though since that model you shared still says it is for causal language modeling, rather than just
Qwen2Model
但这还不够,因为您共享的模型仍然表示它用于因果语言建模,而不仅仅是Qwen2Model
I test the vllm openai server. And I found embeddings and normal chat api cannot use together.
When I use embeddings model start server, I cannot use the same chat model
the same bug~
🚀 The feature, motivation and pitch
in the Mteb leaderboard, the current best embedding model is
Alibaba-NLP/gte-Qwen2-7B-instruct
. However, using the embedding endpoint on it returns the following error currently:can we get support for other embedding models other than E5-mistral?
Alternatives
No response
Additional context
No response