Closed biandh closed 4 years ago
Hi @biandh, We have updated README.md with the answers.
finding that you got magic improvement by finetuning in valid, and got 0.87-0.88 from single model(For VisualBERT, it was around 0.69. As for MCAN, it was around 0.7)....by the way, how did it improve by blending your 69 models?
This huge performance boost is not from post-finetuning, but the post-processing part. We found out that the candidate products provided in the dataset have some bias. That is, if one item occurs in multiple candidate pools it is more likely this item is not the actual answer but just some bias.
thx,amazing magic !
You are so good! I read the READM.md carefully, but I was still confused, mainly in the following aspects:
Can you give a simple example of the negative sample sampling method?
In this competition, 69 models have been trained based on Mcan and Visual Bert methods. Do these 69 models have any differences, such as parameters, training samples, etc?
Before post-processing, a single model based on Mcan or visual Bert is used to evalute the NDCG@5 on valid.tsv. How much can be achieved?
In the post-processing stage, the valid set is used to train the model. How to evaluate the model?
After post-processing, how many Score can a single model achieve in testA?
Looking forward to your reply ^_^