You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Possibly the problem is in the gguf itself? Even so, it's weird that using the "official" gguf results in incorrect tokenization output from the template applied.
Now, you could just always use --chat-template phi3. But my expectation is the phi3 chat template should automatically be picked up by the detection heuristic, when using the canonical/official Phi-3 models, since they purport to support phi3.
The text was updated successfully, but these errors were encountered:
When using phi-3 without the option
--chat-template phi3
, the tokenization is incorrect.For example, if I do use
--chat-template phi3
, here is the log output when I send the message "hi":actually the extra space after <|user|> is concerning, it should be a newline, but maybe that's just an artifact of how the log message is formatted.
But here's what happens when the
--chat-template phi3
is omitted:See how it uses <|endoftext|> (wrong) instead of <|end|> (correct) which causes really bad generation.
I am using the gguf straight from Microsoft, so I guess it is as official as it gets:
https://huggingface.co/microsoft/Phi-3-mini-4k-instruct-gguf
Possibly the problem is in the gguf itself? Even so, it's weird that using the "official" gguf results in incorrect tokenization output from the template applied.
Now, you could just always use
--chat-template phi3
. But my expectation is thephi3
chat template should automatically be picked up by the detection heuristic, when using the canonical/official Phi-3 models, since they purport to support phi3.The text was updated successfully, but these errors were encountered: