-
Notifications
You must be signed in to change notification settings - Fork 1.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How does SFTTrainer handle instruction formatted datasets when a tokenizer has no chat_template? #1233
Comments
Hi @JohnGiorgi |
TLDR is that if you need to use a chat dataset you need to use a model that supports chat templating - if you want to use that model I think you can clone it and manually add a chat template in that model |
Hey @JohnGiorgi, There are fallback templates in
If you are not defining a |
Thanks all! I understand now better and agree that it is probably best to explicitly define a template. |
#1242 might be interesting for you. |
Hi! I am interested in using the
SFTTrainer
for instruction-tuning. Following the docs, I can see that I can provided examples in the following format to have the trainer format things for me:The docs also say:
My question and confusion is, what does the trainer do if the tokenizer has no
chat_template
, as is the case with the base llama model?The text was updated successfully, but these errors were encountered: