Why pytorch-lightning cost more gpu-memory than pytorch? #6653
Unanswered
dalek-who
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 2 comments
-
Also, the most confusion is that the loss curves and final performance between pl and pure pytorch are very different, but I check during the first several steps, loss are the same. Whether GPUAccelerator do something silently and after many steps, the gradients and weights gets more and more different? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Hi @dalek-who , |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-

This is my-gpu usage, The up is pytorch-lightning and the down is pure pytorch, with same model, same batch_size, same data and same data-order, but pytorch-lightning use much more gpu-memory.I use only one GPU, and here's my trainer:
and during training, I check
trainer.accelerator
ispytorch_lightning.accelerators.gpu.GPUAccelerator
, is this accelerator same as pure pytorch with single GPU?Beta Was this translation helpful? Give feedback.
All reactions