Closed boringwar closed 4 years ago
I am new to pytorch DistributedDataParallel (DDP), and not clear about the shuffleBN process.
In the code, you first do _concat_allgather(), and then broadcast a random indexes to every devices from src=0.
Here is my question: Is only device 0 broadcasting? Does other devices doing __batch_shuffleddp()?
After some search, I think I have understood this point.
I am new to pytorch DistributedDataParallel (DDP), and not clear about the shuffleBN process.
In the code, you first do _concat_allgather(), and then broadcast a random indexes to every devices from src=0.
Here is my question: Is only device 0 broadcasting? Does other devices doing __batch_shuffleddp()?