vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
30.43k stars 4.6k forks source link

deploying embedding model in same way as LLM #6498

Open riyajatar37003 opened 4 months ago

riyajatar37003 commented 4 months ago

🚀 The feature, motivation and pitch

can we deploy the embedding model such as bge,e5 etc. using vllm? if yes could u point to the resources for it.

Alternatives

No response

Additional context

No response

ghost commented 4 months ago

see https://github.com/vllm-project/vllm/pull/6282#issuecomment-2230018935

yuanzhiyong1999 commented 3 months ago

🚀 The feature, motivation and pitch

can we deploy the embedding model such as bge,e5 etc. using vllm? if yes could u point to the resources for it.

Alternatives

No response

Additional context

No response

Have you solved it?

ghost commented 3 months ago

No. looking for help

github-actions[bot] commented 3 weeks ago

This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you!