Using Mixed Precision but calculate loss using Full Precision #9132
Unanswered
JoakimHaurum
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment
-
Dear @JoakimHaurum, Yes, you can use torch.cuda.autocast(enabled=False) around the forward of your loss computation. Also, we just merge support for bfloat16 which is more stable than float16. It can be used with Best, |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I am currently training a network for multi label classification.
For this I am using the Asymmetric Loss from Alibaba (https://github.com/Alibaba-MIIL/ASL)
However, when I train with mixed precision the loss goes to NaN.
This is a known problem (Alibaba-MIIL/ASL#53) and has been circumvented by calculating the loss in full precision 32-bit, while during the rest of the forward pass in mixed precision.
Is this possible with pytorch lightning or should I refactor the loss implementation?
Beta Was this translation helpful? Give feedback.
All reactions