-
Notifications
You must be signed in to change notification settings - Fork 134
Sample fine-tuning environment #157
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Also improve default checkpointing strategy.
bf16_mode and bring back autocast
rchan26
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @wesselb - this looks great! Just some small comments. The main one is about ensuring that we definitely use the right .venv when in the image. Once we've made sure that is the case, I think we can merge this in :)
Co-authored-by: Ryan Chan <rchan@turing.ac.uk>
…t/aurora into wesselb/fine-tuning-adjustments
rchan26
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks good to me! thanks again @wesselb !
|
Thanks for the review, @rchan26 :) |
bf16_modeand brings backautocast(Closes Illegal Memory Access with Mixed Precision #121.).attn_drop_ratein the backbone. This will cause stochasticity if the model wasn't.eval()d.I've verified that the instructions work on a fresh Azure VM.