Closed Zi-hao-Wei closed 2 years ago
Hi @Zi-hao-Wei! Thanks for your PR! I know adding accumlate_iter
may help us pre-train VideoMAE with fewer GPUs, but it could lead to lower performance (less than 1%). We will consider supporting 8 GPUs pre-training until we find a more stable way.
Looking forward to seeing the official 8 GPU pre-training code.
I add a accumlate_iter in the pretrain process. This may help we have a same effecitve batchsize as the original one VideoMAE used with less GPUs. I should say 64 GPUs are not usually affordable.