-
Given the solution for exercise 7.1, I would like to ask if this is even a good idea to change the prompt template style without adjusting the tokenizer along the way. In this exercise, we are using the GPT-2 tokenizer but experimenting with Alpaca style and Phi-3 style prompt template styles. But both Alpaca and Phi-3 use the Llama-based tokenizer (that uses different special tokens and a different vocab). Would that explain why the Alpaca and Phi-3 instruction finetuned models perform so badly (assessed with Ollama using Llama 3)? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
You are absolutely correct there. The Phi-3 template would work, but special tokens like I am adding some bonus materials via #496 to address that point :) |
Beta Was this translation helpful? Give feedback.
You are absolutely correct there. The Phi-3 template would work, but special tokens like
<|user|>
would be encoded inefficiently (5 tokens instead of 1) etc.I am adding some bonus materials via #496 to address that point :)