vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
23.05k stars 3.27k forks source link

[Bug]: qwen1.5-32b-chat no response #5957

Open linpan opened 2 weeks ago

linpan commented 2 weeks ago

Your current environment

vllm 0.5.0.post

🐛 Describe the bug

vllm 0.5.0.post

transformers

linpan commented 2 weeks ago

CleanShot 2024-06-28 at 19 41 43