Closed JulioZhao97 closed 1 week ago
Hi, this actually worked as expected. by default, vllm
uses the --model (Name or path of the huggingface model to use) as the model name; if you would like to change the model name, set "--served-model-name". see https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#command-line-arguments-for-the-server for more info.
Hi, this actually worked as expected. by default,
vllm
uses the --model (Name or path of the huggingface model to use) as the model name; if you would like to change the model name, set "--served-model-name". see https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html#command-line-arguments-for-the-server for more info.
Thanks for reply, so why I encounter this bug? What version of vllm
are you currently using?
vllm serve Qwen/Qwen2.5-7B-Instruct
, the model name is Qwen/Qwen2.5-7B-Instruct
.vllm serve Qwen/Qwen2.5-7B-Instruct --served-model-name Qwen2.5-7B-Instruct
, the model name is Qwen2.5-7B-Instruct
.
- if you use
vllm serve Qwen/Qwen2.5-7B-Instruct
, the model name isQwen/Qwen2.5-7B-Instruct
.- if you use
vllm serve Qwen/Qwen2.5-7B-Instruct --served-model-name Qwen2.5-7B-Instruct
, the model name isQwen2.5-7B-Instruct
.
Got it! Thanks for reply.
Model Series
Qwen2.5
What are the models used?
Qwen2.5-7B-Instruct
What is the scenario where the problem happened?
vllm deployment with Qwen2.5-7B-Instruct
Is this a known issue?
Information about environment
vllm==0.6.2
Log output
Description
Steps to reproduce
This happens to Qwen2.5-7B-Instruct The problem can be reproduced with the following steps:
vllm serve Qwen/Qwen2.5-7B-Instruct
Expected results
Retuned a model does not found error:
Attempts to fix
change
model="Qwen2.5-7B-Instruct",
tomodel="Qwen/Qwen2.5-7B-Instruct",
fix the bug