Closed samiwilf closed 1 year ago
This pull request was exported from Phabricator. Differential Revision: D41174060
This pull request was exported from Phabricator. Differential Revision: D41174060
This pull request was exported from Phabricator. Differential Revision: D41174060
This pull request was exported from Phabricator. Differential Revision: D41174060
This pull request was exported from Phabricator. Differential Revision: D41174060
This pull request was exported from Phabricator. Differential Revision: D41174060
Summary: Remove dependence on torch.distributed.algorithms.join. Instead size batches such that all ranks always have the same num_batches. This is possible by increasing batch sizes by 1 sample when necessary to keep num_batches equal across ranks.
Differential Revision: D41174060
LaMa Project: L1141030