Open luluforever opened 5 years ago
还有一个问题是,如何设置预训练的epoch啊?
Sorry for the late response. When you pre-train on multiple machines, make sure that your pytorch versions are the same.
In fact, most pre-training works report their training steps instead of epochs. By now UER doesn't include epoch option.
AttributeError: module 'torch.distributed' has no attribute 'init_process_group'