Open songmzhang opened 3 years ago
The same problem encountered. Hope to be addressed as soon as possible.
Thank you for raising the concern. We do not use a multi-GPU setup and conducted all the experiments utilizing a single GPU (RTX 3090 Ti ). So, our code isn't focused on distributed functions in the fairseq library.
Due to time constraints, I cannot modify the code now but we will try to add this functionality to our code later. However, you may change some parts of the code in order to use multiple GPUs.
Hi there, a little bug occured when I try to use two or more gpus to train the model with your code. To exclude the problem caused by environment, we follow the same environment as you suggested, but the bug still exists.
Here's the error information:
Have you met this problem before? Or have you tried to train your model with multiple gpus?