So I have a model where the outputs of level embedding are being fed to a deep learning model. I'm trying to train the model on multiple GPUs using DDP but face the following error:
RuntimeError: Modules with uninitialized parameters can't be used withDistributedDataParallel. Run a dummy forward pass to correctly initialize the modules
The error occurs only when using Distributed Data Parallel (DDP). Works fine when using Data Parallel (DP).
So I have a model where the outputs of level embedding are being fed to a deep learning model. I'm trying to train the model on multiple GPUs using DDP but face the following error:
RuntimeError: Modules with uninitialized parameters can't be used with
DistributedDataParallel. Run a dummy forward pass to correctly initialize the modules
The error occurs only when using Distributed Data Parallel (DDP). Works fine when using Data Parallel (DP).