The code is waiting infinitely on distributed.barrier(rank) statement in the runs.py file. Please suggest a way to run it in multi-gpu setting. I am running colbertv1 branch code. My OS and pytorch related details are as follows:
**1. pytorch 1.12.0 py3.7_cuda11.3_cudnn8.3.2_0 pytorch
OS is Ubuntu**
For single GPU setting the code works fine for me.
Hi,
While running the multi-GPU training with the below command:
`CUDA_VISIBLE_DEVICES="0,1" python -m torch.distributed.launch --nproc_per_node=2 -m colbert.train \
The code is waiting infinitely on distributed.barrier(rank) statement in the runs.py file. Please suggest a way to run it in multi-gpu setting. I am running colbertv1 branch code. My OS and pytorch related details are as follows: **1. pytorch 1.12.0 py3.7_cuda11.3_cudnn8.3.2_0 pytorch
For single GPU setting the code works fine for me.