vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
26.76k stars 3.92k forks source link

[Bug]: vllm:num_requests_waiting is not being published at /metrics endpoint #7918

Open IshmeetMehta opened 2 weeks ago

IshmeetMehta commented 2 weeks ago

🐛 Describe the bug

Data for vllm:num_requests_waiting is missing.

vllm:num_requests_waiting is not being published at /metrics endpoint

docker image for vllm : vllm-openai:v0.5.3.post1

# HELP vllm:num_requests_waiting Number of requests waiting to be processed.
# TYPE vllm:num_requests_waiting gauge
vllm:num_requests_waiting{model_name="/data/models/model-gemma2-a100/experiment-it1"} 0.0

Before submitting a new issue...

MahouShoujoMivutilde commented 2 weeks ago

Do NOT run it.

It's malware that'll steal your account if executed, to spread further spamming the same message elsewhere, like happened to this person.

There are a lot of comments like that

https://github.com/search?q=is%3Aissue+%22In+the+installer+menu%2C+select+%5C%22gcc.%5C%22%22+AND+%22password%3A+changeme%22&type=issues&s=created&o=desc

(and this is how I found this issue)