vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
25.5k stars 3.69k forks source link

[Feature]: support lora such as qwen-7b and qwen1.5 #4677

Open kynow2 opened 3 months ago

kynow2 commented 3 months ago

🚀 The feature, motivation and pitch

we had trained a lot of lora with qwen-7b ,if vllm support qwen-7b not only qwen1.5 ,that will be better,thanks

Alternatives

No response

Additional context

No response

1149722739 commented 3 months ago

same request