Closed faceless-rex closed 2 years ago
I follow the paper setting, using a batch size of 16 and a audio length of 16000. But 1 GPU is enough for me to train the model. Why does the original paper use 8 GPUs?
You can use as many GPUs to train as you'd like. Multi-GPU training may be faster than single-GPU training.
I follow the paper setting, using a batch size of 16 and a audio length of 16000. But 1 GPU is enough for me to train the model. Why does the original paper use 8 GPUs?