-
Notifications
You must be signed in to change notification settings - Fork 284
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Exception: Current loss scale already at minimum - cannot decrease scale anymore #280
Comments
You can use fp32 to train the model or decrease the batch size to 1. |
Thanks. I will try it. :) |
You can also try using bfloat16 by replacing |
OK, I have already resolved this problem. Thanks :) |
This will not affect the results much. In fact, bf16 will be more stable in training than fp16 and will not suffer from the "Current loss scale already at minimum" problem. |
I see. Many thanks!
|
Thank you for sharing your codes.
When running gpt2/kd/kd_medium.sh on 2*3090, the program encountered this error. What should I do, such as adjusting the learning rate?
The text was updated successfully, but these errors were encountered: