-
-
Notifications
You must be signed in to change notification settings - Fork 5.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Missing prometheus metrics in 0.3.0
#2850
Comments
Just realised the image I'm pulling for the deployment uses vllm/engine/metrics.py from Would it be possible to push another image version to docker hub with the updates? |
I think a new release will be pushed soon -> #2859 |
Hi, @SamComber When I do a
|
@grandiose-pizza did you start your server with |
@hmellor , no It is set to false while starting:
Do I need to add anything to this line?
|
Also it's worth noting that what you're seeing is different because the original screenshot was taken before we switched from |
Could you please share what is expected while using prometheus_client instead? Is it different than the comment above? |
Changing Prometheus client packages only changes the non- The |
It is quite strange. Trying to figure how to obtain the stats like here: Line 20 in 563c1d7
|
I have encountered the same issue as well. If you have resolved it, Pleasel let me know. Thank you. |
I am facing same issue |
is there any update or workaround for this issue? |
Seeing same thing, only basic stats in metrics, no usage, and promethus is not being populated.
I think maybe broken in 0.5.4. On SAME host system also running 0.5.4, just different model, I get more stuff:
is it possible that some models do not support those other metrics? |
@hmellor Why was this issue closed as not planned? It seems like clearly a bug for a useful thing. |
|
@pseudotensor Annoyingly, "not planned" can mean many things (why we can't specify which thing, I don't know), but this was closed as stale originally. |
No problem, it's all working in main. Thanks! |
First of all, thanks for the great open source library!
The docs promise a few more additional metrics that I'm not seeing in vLLM 0.3.0, have these been removed? I.e. if I hit
/metrics
of the OpenAI API server for a deployed model... you'll see novllm:time_to_first_token_seconds
orvllm:time_per_output_token_seconds
orvllm:e2e_request_latency_seconds
The text was updated successfully, but these errors were encountered: