When I train the code, the multi-GPU allocation stops at the following location:
"
Resuming checkpoint None, exp_version=None
initializing ddp: GLOBAL_RANK: 1, MEMBER: 2/5
Resuming checkpoint None, exp_version=None
initializing ddp: GLOBAL_RANK: 4, MEMBER: 5/5
When I train the code, the multi-GPU allocation stops at the following location: " Resuming checkpoint None, exp_version=None initializing ddp: GLOBAL_RANK: 1, MEMBER: 2/5 Resuming checkpoint None, exp_version=None initializing ddp: GLOBAL_RANK: 4, MEMBER: 5/5
" Is this training? Or something wrong ?