Closed Maxusmusti closed 8 months ago
Additionally, I can set the value of early_stopping
to any random value, and it will throw no error, so it seems like it isn't even being passed to the SamplingParameters
Upon further inspection, it looks like it has just been left out of the completion/chat-completion apis, possibly an oversight?: https://github.com/vllm-project/vllm/blob/main/vllm/entrypoints/openai/protocol.py#L56
Opened a PR with a potential quick 4-line fix, let me know if it looks like anything is missing! https://github.com/vllm-project/vllm/pull/2939
It's not an oversight originally because it is not part of the official API https://platform.openai.com/docs/api-reference/chat/create
But it does seem needed. Thank you for your PR
use_beam_search
and length_penalty
also aren't part of the official API, it goes along with those I guess
Closed by #2939
While using
v0.3.1
,early_stopping
will not toggle toTrue
due to an omission in the protocol definition (see below comments). I am prompting like this:and this is what I get on server side:
Everything else I set is there, but the early stopping isn't, even though none of my other options should be incompatible: https://github.com/vllm-project/vllm/blob/264017a2bf030f060ebad91eb9be9b4e0033edb9/vllm/sampling_params.py#L106