Open xialeiliu opened 5 years ago
Is there a parallel data loader available?
you might want to try increasing the batchsize? this will relax the data loading issue quite much.
I already used the largest batch size I can in a 8-gpu machine. that's might not be the solution.
I found that the gpus are waiting for cpus most time, and cpu usage is very low even with num_workers = 8, that's might slow down the training process. Do you have any idea how to improve the speed of data loader?