Can you please tell, what annotation files are used for calculating the visual features of OKVQA dataset or train and validation split?
Also, please let me know do we need to calculate image features for whole coco dataset in order to calculate okvqa dataset's features? As, okvqa dataset only has 15k images but i am unable to find only those images anywhere.
❓ Questions and Help
The first step of KRISP model before VQA is to calculate visual features of OKVQA dataset. For this Bottom-up features are used probably from https://github.com/peteanderson80/bottom-up-attention .
Can you please tell, what annotation files are used for calculating the visual features of OKVQA dataset or train and validation split?
Also, please let me know do we need to calculate image features for whole coco dataset in order to calculate okvqa dataset's features? As, okvqa dataset only has 15k images but i am unable to find only those images anywhere.