-
-
Notifications
You must be signed in to change notification settings - Fork 5.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Unrecognized keys in rope_scaling
for 'rope_type'='linear': {'type'}
#6897
Comments
Please update your version of |
I have already used the latest version of 4.43.3 |
Can you post the full stack trace so we can investigate? |
According to the log, the rope_scaling has 3 keys: type, factor, rope_type, however, not sure which code check it.
|
I think this is a warning message that can be safely ignored (see huggingface/transformers#32182). The model still works, right? |
Ok, now I can confirm the rope scaling is used correctly. The model works well, thanks for your help. |
Your current environment
🐛 Describe the bug
When using vllm 0.5.3 and 0.5.3.post1 to deploy deepseek coder 6.7b, which has a rope_scaling configuration in the config.json, a message appears "Unrecognized keys in
rope_scaling
for 'rope_type'='linear': {'type'}", not sure which script raised it.serve code:
/usr/bin/python3 -m vllm.entrypoints.openai.api_server$SERVER_PATH/"$ {MODEL_NAME}" $SERVER_PATH/"$ {MODEL_NAME}"
--host ::
--port "${PORT0}"
--model
--served-model-name $SERVED_MODEL_NAME
--tensor-parallel-size "${GPU_NUM}"
--tokenizer
--max-model-len $MAX_LENGTH
--gpu-memory-utilization 0.9
--speculative-model "[ngram]"
--ngram-prompt-lookup-max 3
--ngram-prompt-lookup-min 1
--num-speculative-tokens 5
--use-v2-block-manager
--enable-prefix-caching
--trust-remote-code
--dtype auto
The text was updated successfully, but these errors were encountered: