Closed jinhyun95 closed 4 years ago
The config projects/hateful_memes/configs/visual_bert/from_coco.yaml
is used only for training as it has checkpoint.resume_pretrained=True
and loads only the bert
part of the model from a coco pretrained visual bert model and the classification weights are not loaded.
For validation and inference when you are loading a finetuned model use this config : projects/hateful_memes/configs/visual_bert/defaults.yaml
So your command should be this :
mmf_run config=projects/hateful_memes/configs/visual_bert/defaults.yaml model=visual_bert dataset=hateful_memes run_type=val checkpoint.resume_zoo=visual_bert.finetuned.hateful_memes.from_coco
@vedanuj Thanks!! Same should be applied for Vilbert, I suppose?
@jinhyun95 Yes, you are right.
Hi @vedanuj I ran the same command that you mentioned for validation:
mmf_run config=projects/hateful_memes/configs/visual_bert/defaults.yaml model=visual_bert dataset=hateful_memes run_type=val checkpoint.resume_zoo=visual_bert.finetuned.hateful_memes.from_coco
However I am not able to exactly reproduce the numbers from the paper.
For the above command for visual bert finteuned on coco I get roc_auc=0.7342
and accuracy=0.6320
, which are slightly lower than those reported in the paper.
For vilbert finetuned with cc I get roc_auc=0.7067
and accuracy=0.6280
, which are slightly higher than those reported in the paper.
Am I missing something?
Hi @purvaten .. the numbers reported in the paper are average of multiple runs with different seeds. So there might be slight differences.
@vedanuj I see, makes sense. Thanks for the clarification!
❓ Questions and Help
When executing the following command to evaluate the pretrained model zoo, the results differ every time, and none of the results shows the results provided in https://arxiv.org/abs/2005.04790.
run 1:
run 2:
run 3:
run 4: