Closed yash0307 closed 2 years ago
Hi @yash0307
Thanks for your interest in our work.
Sometimes, the multi-label training is not stable enough. You may set multi_label=0
in the config or pass --opts data.text_aug.multi_label=0
when you launch training.
It should yield 1% lower mIoU than what we reported.
I will try that. Thank you for the quick response.
Dear Authors,
I tried training the GroupViT model on GCC + YFCC datasets using the
group_vit_gcc_yfcc_30e.yml
config file and with a batch size of 2048 (256x8). The results on PASCAL VOC after 30 epochs of training is roughly 5% mIoU lower (absolute) than what is reported in the paper. Additionally, I tried training with gradient accumulation with 2 steps, which did not give any improvement. Do you have any suggestions on what can cause the lower performance?Thank you.