Closed Euphoria16 closed 3 years ago
Yep, I think some easy way to enable it is to just skip that line -- because it performing an all-gather operation from all gpus, and since in the non-distributed training it only has one gpu, so this line is not needed.
I found if I didn't use distributed training, i.e. set the --multiprocessing-distributed=False and use single GPU, there seems to be no problems in main_moco.py with
However, this error occurred when training started
This error can be traced back to
and
This error is caused by computation of contrastive_loss, which still relies on distributed training. So I wonder if the non-distributed training is not supported even if we set multiprocessing-distributed=False.