-
-
Notifications
You must be signed in to change notification settings - Fork 5.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: benchmark_serving model_id bug for lmdeploy #4001
Comments
I wouldn't call this a bug because unlike other inference backends that use the huggingface model id also as the default model name for the API server, the model name for
The benchmark script already provides the flexibility to allow users to specify
I can make a PR to make this clearer if that helps. |
@ywang96 In order to make this benchmark run as expected, perhaps we can add a parameter similar to |
Wouldn't this work for
Perhaps I can make the intention of |
make sense |
It's ok. |
Your current environment
🐛 Describe the bug
Hi @ywang96 Currently there is a small issue in benchmarks/backend_request_func when benchmark LMDeploy with Llama-2-13b-chat-hf.
I need to change request_func_input.model to
llama2
vllm/benchmarks/backend_request_func.py
Line 222 in f3d0bf7
After manual modification, testing can be conducted, and the correct result is:
Otherwise, the test result is incorrect because the model name was not correctly matched.
The text was updated successfully, but these errors were encountered: