Open priyamtejaswin opened 2 years ago
Hi @priyamtejaswin ,
Sorry for the confusion. TAP-OCRCC uses 0/12 layers instead of 3/4 layers (detailed in paper Table 5). Thus we need to update the layer number in the config file. I'll try to update the config file, before that, you could update the layer number to 0/12 (3->0, 4->12) here and see if it solves the problem.
https://github.com/microsoft/TAP/blob/352891f93c75ac5d6b9ba141bbe831477dcdd807/configs/vqa/m4c_textvqa/tap_refine.yml#L57 https://github.com/microsoft/TAP/blob/352891f93c75ac5d6b9ba141bbe831477dcdd807/configs/vqa/m4c_textvqa/tap_refine.yml#L66
Hello devs,
Thank you for publishing this work, and for sharing these resources!
I was trying to run the evaluation code for TextVQA that is mentioned in the README. I can successfully run the following using textvqa_tap_base_best.ckpt
I believe this returns the results without the addition (OCR-CC) dataset. I think that checkpoint is saved under
save/finetuned/textvqa_tap_ocrcc_best.ckpt
However, when I use the
ocrcc
checkpoint, it fails while loading the checkpoint...Do I need to change the
--model
argument passed torun.py
? At the moment it is--model m4c_split
. This is the command to reproduce the above error: