-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Train] Add example of fine-tuning Llama-2 on Intel Gaudi #44667
Conversation
2886045
to
5b74dfe
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
let docs reviewer know when this is ready for review.
e1bd5a9
to
fb0028f
Compare
570f385
to
f5f3268
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for the contributing these two high-quality examples @harborn ! Great demonstration of using LoRA and deepspeed ZERO-3 on Gaudi + Ray for fine-tuning.
Left some comments.
f76ccee
to
1be2370
Compare
1be2370
to
5bca190
Compare
@justinvyu Can you take a look and merge it? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The examples look good! Just a few requests:
- Clear the cell outputs, and just put a mock markdown cell with the important output info. For example, just this information:
train_result = TrainOutput(global_step=62, training_loss=1.500297857869056, metrics={'train_runtime': 93.3311, 'train_samples_per_second': 71.042, 'train_steps_per_second': 2.222, 'total_flos': 4.02963202792489e+16, 'train_loss': 1.500297857869056, 'epoch': 2.0, 'memory_allocated (GB)': 34.51, 'max_memory_allocated (GB)': 78.72, 'total_memory_available (GB)': 94.62})
- Is it possible to merge these two notebooks so that I can just flip a flag if I want to use deepspeed? Most of the logic is identical, just some extra configs.
- (Just a question, not blocking) Should we also allow full parameter finetuning instead of always using lora?
429cb50
to
98eb4df
Compare
Hi, Updated according your comments:
|
Seems that @harborn addressed the comments. @justinvyu could you take a look again? |
@justinvyu please take a look again! Thanks. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for making the changes! One last comment then we can merge.
Fixed. |
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
Signed-off-by: Wu, Gangsheng <[email protected]>
…t#44667) Adds an example for fine-tuning Llama-2-7b/70b on multiple HPUs. --------- Signed-off-by: Wu, Gangsheng <[email protected]> Signed-off-by: Ryan O'Leary <[email protected]>
…t#44667) Adds an example for fine-tuning Llama-2-7b/70b on multiple HPUs. --------- Signed-off-by: Wu, Gangsheng <[email protected]> Signed-off-by: Ryan O'Leary <[email protected]>
…t#44667) Adds an example for fine-tuning Llama-2-7b/70b on multiple HPUs. --------- Signed-off-by: Wu, Gangsheng <[email protected]>
Why are these changes needed?
To leverage the potential of Intel Gaudi accelerator, we extend Ray Train's capabilities by adding support for Intel Gaudi (HPU) hardware. This PR include an example for fine-tuning Llama-2-7b on multi HPUs.
Related issue number
Checks
git commit -s
) in this PR.scripts/format.sh
to lint the changes in this PR.method in Tune, I've added it in
doc/source/tune/api/
under thecorresponding
.rst
file.