Closed ma787639046 closed 1 year ago
Hi, thanks for your excellent work.
As described in the Experiment Settings Section, the pretraining of retromae with enhanced decoding was finished on 8*A100 GPUs. Could you please tell us how many hours you took to complete the pretraining?
Thanks.
Hi, it takes about 2 days on wiki and bookcorpus dataset, and no more than 1 day on msmarco dataset.
Thanks a lot!
Hi, thanks for your excellent work.
As described in the Experiment Settings Section, the pretraining of retromae with enhanced decoding was finished on 8*A100 GPUs. Could you please tell us how many hours you took to complete the pretraining?
Thanks.