runpod-workers / worker-vllm

The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
MIT License
213 stars 82 forks source link

OpenAI Error: Not returning full output #58

Closed Mr-Nobody1 closed 5 months ago

Mr-Nobody1 commented 5 months ago

image

alpayariyak commented 5 months ago

Specify the max_tokens parameter in your input