-
Notifications
You must be signed in to change notification settings - Fork 889
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Suggestion] Instruction Fine-Tuning - SFT Module #60
Comments
Thanks. That's a nice suggestion. Are you interested in opening a PR? |
I can give it a try, however, I am not able to do it in a timely manner. I will be available starting from the 18th. |
No worries @ArlindKadra . Come back when you're ready and check if it still needs doing. Thanks for the issue. |
Wanted to add few of my queries here, instead of creating a new issue. (I hope this is ok)
Tried to get answers for these queries in SFT couldn't get answers. Maybe adding bit more detail in "The Finetuning Process" sub-section of the Supervised Fine tuning page would be helpful. |
Thanks for taking the time into developing this interesting course. I wanted to suggest that regarding the SFT module in Chapter 1, the
bigcode/the-stack-smol
dataset seems to break the flow a bit. Since it is not an instruction-tuning dataset, but more a domain-specific dataset with no instruction following. As such, it does not have the question/answering pairs.Based on that, if you train on the dataset, the response to the prompt is the same as before. Maybe switching it up to the
openai/gsm8k
dataset, or something similar? That way one would still have to prepare the dataset before feeding it to the SFTTrainer.The text was updated successfully, but these errors were encountered: