-
Notifications
You must be signed in to change notification settings - Fork 136
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
fine-tune HKUNLP/instructor-embedding #74
Comments
I only have training data with format: sentence1, sentence2, label |
Also, when I trying to train using train.py, with "--fp16 True --gradient_accumulation_steps 3", I got out of GPU memory. I was using A100 40G. why training this model takes this much GPU memory. could you tell me the GPU hardware you used to train this model? |
btw, this model can be trained only when per_device_train_batch_size is set to 2 |
@Atlantic8 , this is an excerpt from the paper:
|
What's your source for this? @Atlantic8 |
Hi, Thanks a lot for your interest in the INSTRUCTOR!
Hope this helps! |
So for custom data, do we need to randomly construct a data format like query=xxx, pos=[], neg=[] before running? |
can we fine-tune using train.py based on the released model hkunlp/instructor-xl? If yes, could you please show me the shell script for training? thanks
The text was updated successfully, but these errors were encountered: