ChenRocks / UNITER

Research code for ECCV 2020 paper "UNITER: UNiversal Image-TExt Representation Learning"
https://arxiv.org/abs/1909.11740
781 stars 109 forks source link

pre-trained model #57

Open yixuan-qiao opened 3 years ago

yixuan-qiao commented 3 years ago

hi, the uniter-large.pt released is trained on in-domain & out-of-domain data or only in-domain data?

yixuan-qiao commented 3 years ago

Hi, Thanks for your excellent work. I am not sure the batchsize in your paper is same as it in the code? In code, 3072 refers to total tokens, corresponding to about real 32 examples each iteration.

a) Maybe 32(real batchsize)4(Grad. Accu) is dominant factor? b) Our V100 machine (16G) can not process the 3072 tokens, so maybe 1024 tokens(about 8 real examples), 8 Gpus, 2(Grad. Accu) is another workable plan? because 324=882 c) Besides, the train-vqa-large-8gpu-adv.json you released can reproduce the best large model in paper result?

We deeply hope to reproduce your best results in our limited resource scenario. Thank a lot. UNITER and VILLA are really valuable work!