I read in the issues that to achieve 54% validation performance on mit-b2, it requires using 4 2080 GPUs.
My question is, how much can distributed training improve the final validation performance? Also, if I only have a single GPU, what should I do or what hyperparameters should I use to achieve results close to those reported in the paper? Thank you!
Hello, I have been training using the A6000 with the following command, the dataset is NYUdepthV2:
The final evaluation results are as follows:
I read in the issues that to achieve 54% validation performance on mit-b2, it requires using 4 2080 GPUs.
My question is, how much can distributed training improve the final validation performance? Also, if I only have a single GPU, what should I do or what hyperparameters should I use to achieve results close to those reported in the paper? Thank you!