You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am currently working with the network you provided, and I have encountered an issue that I need assistance with. Specifically, when training the model using half-precision (FP16), I consistently experience numerical overflow errors. However, when using full precision (FP32), the training proceeds without any issues. I have tried adjusting various settings, but the problem persists.
Could you kindly offer any suggestions or solutions to address this issue?
Thank you very much for your valuable work, and I look forward to your response.
Best regards,
Wenqiang
The text was updated successfully, but these errors were encountered:
Hi,
We also didn't manage to train with FP16, due to overflows. Please try bfloat16 instead, with which we had more luck. Unfortunately, it's only available on relatively recent NVIDIA GPUs.
Pim
Dear Author,
I hope this message finds you well.
I am currently working with the network you provided, and I have encountered an issue that I need assistance with. Specifically, when training the model using half-precision (FP16), I consistently experience numerical overflow errors. However, when using full precision (FP32), the training proceeds without any issues. I have tried adjusting various settings, but the problem persists.
Could you kindly offer any suggestions or solutions to address this issue?
Thank you very much for your valuable work, and I look forward to your response.
Best regards,
Wenqiang
The text was updated successfully, but these errors were encountered: