Alibaba-MIIL / ML_Decoder

Official PyTorch implementation of "ML-Decoder: Scalable and Versatile Classification Head" (2021)
MIT License
315 stars 52 forks source link

pretraining for ML_Decoder’ backbone and openimagev6 dataset #12

Closed sorrowyn closed 2 years ago

sorrowyn commented 2 years ago

This is a very Interesting work! May I ask you two questions? Q1: You used Open Images pretraining for ML_Decoder’ backbone to get tresnet_l_pretrain_ml_decoder. parser.add_argument('--model-path', default='https://miil-public-eu.oss-eu-central-1.aliyuncs.com/model-zoo/ML_Decoder/tresnet_l_pretrain_ml_decoder.pth', type=str) 图片

Q2:Will you release the data loading file for openimage?

mrT23 commented 2 years ago

Q1: the pretrain weights we uploaded are indeed from open images ( https://miil-public-eu.oss-eu-central-1.aliyuncs.com/model-zoo/ML_Decoder/tresnet_l_pretrain_ml_decoder.pth ) using them, you can fully reproduce the article results.

Q2: we share our open images dataset variant in: https://github.com/Alibaba-MIIL/PartialLabelingCSL/blob/main/OpenImages.md i am not sure what you mean by " data loading file ". at this moment, we are not providing a specific dataloader for open images. but you can easily write one yourself.

sorrowyn commented 2 years ago

Thank you for your sincere reply! Now I'm using the penImage dataset that you provided.However, it seems to be a missing JSON file containing labels to the index. More important. Happy New Year!

sorrowyn commented 2 years ago

Q1: the pretrain weights we uploaded are indeed from open images ( https://miil-public-eu.oss-eu-central-1.aliyuncs.com/model-zoo/ML_Decoder/tresnet_l_pretrain_ml_decoder.pth ) using them, you can fully reproduce the article results.

Q2: we share our open images dataset variant in: https://github.com/Alibaba-MIIL/PartialLabelingCSL/blob/main/OpenImages.md i am not sure what you mean by " data loading file ". at this moment, we are not providing a specific dataloader for open images. but you can easily write one yourself.

Thank you for your sincere reply! Now I'm using the penImage dataset that you provided.However, it seems to be a missing JSON file containing labels to the index. More important. Happy New Year!

kprokofi commented 2 years ago

Hello, thanks for your work. I have another question related to OI. In your article as well in ASL article you wrote that the number of training images is 9 mln, but in your link for downloading dataset 5662868 Isn't it a mistake?

mrT23 commented 2 years ago

Q1: the pretrain weights we uploaded are indeed from open images ( https://miil-public-eu.oss-eu-central-1.aliyuncs.com/model-zoo/ML_Decoder/tresnet_l_pretrain_ml_decoder.pth ) using them, you can fully reproduce the article results. Q2: we share our open images dataset variant in: https://github.com/Alibaba-MIIL/PartialLabelingCSL/blob/main/OpenImages.md i am not sure what you mean by " data loading file ". at this moment, we are not providing a specific dataloader for open images. but you can easily write one yourself.

Thank you for your sincere reply! Now I'm using the penImage dataset that you provided.However, it seems to be a missing JSON file containing labels to the index. More important. Happy New Year!

1) Notice that for pre-training or score calculation, you don't need the actual class names, mid description is sufficient 2) I think the file: https://miil-public-eu.oss-eu-central-1.aliyuncs.com/public/Open_Images_V6/mid_to_classes.pth (which appears in https://github.com/Alibaba-MIIL/PartialLabelingCSL/blob/main/OpenImages.md) contains the conversion between mid to class names

mrT23 commented 2 years ago

5662868

The direct flicker links for downloading OpenImages dataset are becoming invalid over the years. In addition, OpenImages also contains different versions (V3,V4,V6).

We downloaded the dataset ~1.5 years ago, and these are the images that were available then. Now there are less. we are the first to try sharing and standardizing the dataset, instead of letting people use fewer and fewer images.

kprokofi commented 2 years ago

5662868

The direct flicker links for downloading OpenImages dataset are becoming invalid over the years. In addition, OpenImages also contains different versions (V3,V4,V6).

We downloaded the dataset ~1.5 years ago, and these are the images that were available then. Now there are less. we are the first to try sharing and standardizing the dataset, instead of letting people use fewer and fewer images.

Thanks! I just wanted to know, that the pre-trained weights for your models and metrics that you reported were obtained using the dataset you shared for public