Using DataParallel wrapper #10396
Unanswered
w2kun
asked this question in
Lightning Trainer API: Trainer, LightningModule, LightningDataModule
Replies: 1 comment 1 reply
-
Dear @w2kun, This is more complicated :) Lightning doesn't support natively hybrid mechanisms right now unless you implement it yourself. Would you mind sharing more details on what you want to achieve. If you are trying to scale your batch_size, you could use DeepSpeed + precision=16 + activation checkpointing and this should enable you to scale to very large batch size. If it is not enough, you could use Best, |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
In my experiment setting, I need multi-gpu training in certain training stage to increase batch size. Is it right that I set gpus flag in Trainer to 1 (I don't want to use any accelerator) and wrap some submodule with DataParallel when multi-gpu training is needed?
Beta Was this translation helpful? Give feedback.
All reactions