Open hongqing1986 opened 2 weeks ago
Perhaps similar to https://github.com/vllm-project/vllm/issues/9769
+1
Can you try again using the latest version? Supposedly it should be fixed since #9826.
Can you try again using the latest version? Supposedly it should be fixed since #9826.
When will 0.6.4 be released?
A release is quite overdue, we're planning to release the next update this week.
Your current environment
vLLM version: v0.6.3.post1
🐛 Describe the bug
In the latest version v0.6.3.post1, when generating long texts (for example, when the number of tokens reaches 21,000), the generated content is basically garbled. Additionally, after verifying, the long text functionality in v0.6.2 works correctly using the qwen2-7b-instruct model. Furthermore, I also tested other models like qwen2.5-72b-instruct, which exhibit the same problem.
Before submitting a new issue...