-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fine-tuning #23
Comments
I have same question. Thank you for reply. |
+1 |
6 similar comments
+1 |
+1 |
+1 |
+1 |
+1 |
+1 |
There's an RLHF harness here: https://github.com/nebuly-ai/nebullvm/tree/main/apps/accelerate/chatllama Instruction fine tuning should also be pretty easy. |
++ |
+1 |
1 similar comment
+1 |
|
+1 |
4 similar comments
+1 |
+1 |
+1 |
+1 |
Found this work from Stanford Alpaca. They read and resave the model(so not dependent on FAIR) and it becomes trainable with the hugging face training framework. https://crfm.stanford.edu/2023/03/13/alpaca.html |
+1 |
We recently shared scripts for finetuning and inference at https://github.com/facebookresearch/llama-recipes |
Is it possible to Fine-tune LLaMA for downstream tasks? If so, how can we do that?
Edit: Reading the other opened issues, I realized that neither the training data nor the pre-trained weights were released. How the code is going to be useful anyway?
The text was updated successfully, but these errors were encountered: