-
-
Notifications
You must be signed in to change notification settings - Fork 5.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model] Add support for 360zhinao #4078
Conversation
@simon-mo can you help us review the code? |
Getting
On
|
This branch works for the vllm 0.4.0 version. I will merge these 2 new refactor: [Core] Refactor model loading code (https://github.com/vllm-project/vllm/pull/4097[)](https://github.com/vllm-project/vllm/commit/69e1d2fb6922b2d388bae41286d8867976cbd6c6) [Core][Refactor] move parallel_utils into vllm/distributed (https://github.com/vllm-project/vllm/pull/3950[)](https://github.com/vllm-project/vllm/commit/63e7176f265be43dcc425f5ab4ab45c90234f5c3) |
finished merge 4097,3950 |
I'm running into the following issues:
|
it is a chat model, we use the chat api.
the result is :
we will add this config to tokenizer_config.json later: |
The tokenizer_config.json should also include the following so it doesn't need to be specified by the client each time. Please let me know once the hf or modelscope version is updated.
|
thanks, we will fix it |
@simon-mo
|
Looks good, please fix lint by running ./format.sh |
OK
|
|
Did you push the changes? |
updated. thanks |
@simon-mo is it ready to merge? |
Add support for 360zhinao model
We released the 360Zhinao model series:
Notable features of our 360Zhinao models are: