Closed harrytrinh96 closed 3 years ago
I created the dataset following two papers:
I used the word2vec from https://github.com/salman-h-khan/PL-ZSD_Release.
word_w2v_withbg_48_17.txt and word_w2v_withbg_65_15.txt are the word2vecs of seen and unseen classes in two splits. "withbg" means with background word2vec for "background" class.
How to generate the label files like instances_val2014_unseen_65_15.json:
For testing label file: a. for zsi task (only needs unseen classes):
i. Load the val label file instances_val2014.json (provided by the official MS-COCO)
ii. Remove all annotations, all categories belong to seen classes (here we just use the test image lists from above paper)
iii. Save the filtered label file (instances_val2014_unseen_48_17.json for 48/17 split and instances_val2014_unseen_65_15.json for 65/15 split)
b. for gzsi task (needs both seen and unseen classes):
i. Load the val label file instances_val2014.json (provided by the official MS-COCO)
ii. Keep all seen and unseen classes (here we just use the test image lists from above paper)
iii. Save the filtered label file (instances_val2014_gzsi_48_17.json for 48/17 split and instances_val2014_gzsi_65_15.json for 65/15 split)
Thank you very much for your answers. I will look into the papers you provided. If i don't have further questions for this topic, I will close this issue myself.
@zhengye1995 Could you please share your code to load instances_val2014_unseen_65_15.json:
which can "remove all images, all annotations, all categories belong to unseen classes and just keep seen classes data"?
@zhengye1995 May I know how u generate vectors for classes with multiple words (e.g. 'traffic light' in COCO)?
Could you please let me know how did you create datasets? I mean:
what are
vocabulary_w2v.txt, word_w2v_withbg_48_17.txt, word_w2v_withbg_65_15.txt
? And how did you create them?Also, How did you create files in
annotations
folders such asword_w2v_withbg_65_15.txt, instances_val2014_gzsi_65_15.txt, and instances_val2014_unseen_65_15.json
? Thank you.