-
Notifications
You must be signed in to change notification settings - Fork 10.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix code typo in llama-cli #8198
Conversation
I don't think it is completely right yet, there are still extra new lines added to the assistant messages randomly. I suspect that at least one issue is that the chat template expects a new line after the |
Hmm, that's strange. My result is pretty consistent (with
|
That case is also fixed for me, but I still see many messages ending in double or triple lines during random chat. |
Probably there is some thing more specific for gemma template (or the model itself). In any case, I'll merge this PR now and have a deeper look on gemma later. |
I see what you mean. In this case, we must either patch the template behavior with |
Fix a small typo that breaks chat template support on
llama-cli -cnv