Open sanyam5 opened 6 years ago
It turns out that not all 4 of my GPU's were free. Removed the occupied GPU's from train.yml and it's working.
However I noticed that you need to include GPU #0 otherwise pytorch reports that the variables are on different GPU's. I suspect that PyTorch puts some variables on GPU #0 by default.
You are correct. I have made some fixes since then, but I haven't push it to the repo yet. I will hopefully do so soon.
Hey I am getting this error when I try to train.
train.yml
Do you know why this could be happening?