-
-
Notifications
You must be signed in to change notification settings - Fork 5.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: cannot load model back due to [does not appear to have a file named config.json] #4416
Comments
Having the same probleme here. Modifing this script Replacing TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF by bartowski/Meta-Llama-3.1-8B-Instruct-GGUF giving this error :
|
same issue, +1 vllm version: v0.5.5 for GGUF model, I do not think config.json is necessary, please tell me if I am wrong |
+1 OSError: /DeepSeek-Coder-V2-Instruct-GGUF does not appear to have a file named config.json. Checkout 'https://huggingface.co//u01/data/analytics/models/DeepSeek-Coder-V2-Instruct-GGUF/tree/None' for available files. |
Alternatively, if there's a way to know what should be added in a config.json, I'm happy to start adding them.. |
@bartowski1182 I tryed to add localy a My config.json contained this: {
"model_type": "cohere",
} I was inspired by what TheBloke was doing here I got another error telling me that VLLm didn't support models of type "cohere". But the first mistake was fixed. It might work with other repo's (aka LLama2) |
It would also be easier if it just expected the "ForCausalLM" format that is in existing config.jsons cause then I could just parse that and throw it in, but I've seen it being "llama" in theblokes models |
This issue has been automatically marked as stale because it has not had any activity within 90 days. It will be automatically closed if no further activity occurs within 30 days. Leave a comment if you feel this issue should remain open. Thank you! |
This issue has been automatically closed due to inactivity. Please feel free to reopen if you feel it is still relevant. Thank you! |
Your current environment
🐛 Describe the bug
hi there,
I fine-tuned the mistral-7b-v0.2 model using the trainer of huggingface https://huggingface.co/docs/trl/v0.8.6/trainer
the training worked well and finally it saved the model, as bellow:
path_to_the_model:
it can successfully be loaded back using
AutoModelForCausalLM.from_pretrained
However, when I try to load it back via vllm, it caused error:
The vllm codes:
neither works.
any solutions ?
The text was updated successfully, but these errors were encountered: