frank-xwang / CLD-UnsupervisedLearning

[CVPR 2021] Code release for "Unsupervised Feature Learning by Cross-Level Instance-Group Discrimination."
MIT License
100 stars 9 forks source link

CLD: Unsupervised Feature Learning by Cross-Level Instance-Group Discrimination.

by Xudong Wang, Ziwei Liu and Stella X. Yu at UC Berkeley / ICSI and NTU.

IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.

For more information, please check: Project Page | PDF | Preprint | BibTex

Updates

[06/08/2021] Training and linear evaluating InfoMin + CLD on ImageNet is supported.

[05/20/2021] Training and linear evaluating MoCo v2 + CLD on ImageNet is supported.

[04/12/2021] Training MoCo + CLD on CIFAR is supported now.

[04/09/2021] Initial Commit. Training NPID + CLD on CIFAR is avaliable now in this repo. We also plan to support MoCo+CLD, BYOL+CLD and InfoMin+CLD.

Requirements

Packages

Dataset Preparation

CIFAR and STL-10 code will download data automatically with the dataloader. For ImageNet, please download the ImageNet-1k dataset from here. Moving validation images to labeled subfolders using the following script is required: valprep.sh. For ImageNet-100, it was firstly used in CMC and contains 100 categories of ImageNet. The category list of ImageNet-100 can be found in data/imagenet100.txt. Please organize ImageNet-100 according to the following structure. For high-correlation dataset Kitchen-HC, it is constructed by extracting objects in their bounding boxes from the multi-view RGB-D Kitchen dataset. Kitchen-HC can be downloaded from this link.

data
├── CIFAR-10
│   └── cifar-10-batches-py
│       ├── batches.meta
│       ├── data_batch_1
│       ├── ...
│       ├── data_batch_5
│       └── test_batch
├── CIFAR-100
│   └── cifar-100-python
│       ├── file.txt~
│       ├── meta
│       ├── train
│       └── test
├── Kitchen-HC
│   ├── train
│   │   ├── n02869837
│   │   ├── ...
│   │   └── n02090622
│   └── test
│       ├── n02869837
│       ├── ...
│       └── n02090622
├── ImageNet-100
│   ├── train
│   │   ├── n02869837
│   │   ├── ...
│   │   └── n02090622
│   └── val
│       ├── n02869837
│       ├── ...
│       └── n02090622
└── ImageNet
    ├── train
    │   ├── n01440764
    │   ├── ...
    │   └── n15075141
    └── val
        ├── n01440764
        ├── ...
        └── n15075141

Training and Evaluation Instructions

CIFAR-10 and CIFAR-100

NPID + CLD

bash scripts/train_cifar10_npid_cld.sh or bash scripts/train_cifar100_npid_cld.sh
Method Projection Head CIFAR-10 CIFAR-100
NPID Linear 80.8 51.6
NPID+CLD (reported) Linear 86.7 57.5
NPID+CLD (reproduced) Linear 86.8 58.8

The model is trained with mixed precision (fp16) by default, it is necessary to install apex if you want to apply mixed precision training. The reproduced result is the average kNN accuracies of 3 runs.

MoCo + CLD

bash scripts/train_cifar10_moco_cld.sh or bash scripts/train_cifar100_moco_cld.sh
Method Projection Head CIFAR-10 CIFAR-100
MoCo Linear 82.1 53.1
MoCo+CLD (reported) Linear 87.5 58.1
MoCo+CLD (reproduced) Linear N/A 59.1
MoCo+CLD (reproduced) NormLinear N/A 59.7

The model is trained with the proposed NormLinear as the projection head by default. Other settings are the same as NPID+CLD.

ImageNet

MoCo v2 + CLD

train

bash scripts/imagenet/train_imagenet_mocov2_cld.sh

linear evaluation

bash scripts/imagenet/test_imagenet_moco_cld.sh
Method Projection Head #epochs Top-1 (%) Models
MoCov2 MLP 200 67.5 -
MoCov2+CLD MLP 200 69.2 link
MoCov2+CLD NormMLP 200 70.0 link

It is necesary to change the DATA_DIR, SAVE_DIR and PRETRAINED_MODEL. All models are pretrained for 200 epochs and evaluated with standard augmentation, linear decay scheduler and SGD optimizer.

InfoMin + CLD

train

bash scripts/imagenet/train_imagenet_infomin_cld.sh

linear evaluation

bash scripts/imagenet/test_imagenet_infomin_cld.sh
Method Projection Head #epochs Top-1 (%) Models
InfoMin MLP 100 67.4 -
CLD MLP 100 69.5 link
CLD NormMLP 100 70.1 link
CLD MLP 200 70.6 link
CLD NormMLP 200 71.5 link

Please change the DATA_DIR and PRETRAINED_MODEL before launching experiments.

How to get support from us?

If you have any general questions, feel free to email us at xdwang at eecs.berkeley.edu. If you have code or implementation-related questions, please feel free to send emails to us or open an issue in this codebase (We recommend that you open an issue in this codebase, because your questions may help others).

Citation

If you find our work inspiring or use our codebase in your research, please cite our work.

@inproceedings{wang2021unsupervised,
  title={Unsupervised feature learning by cross-level instance-group discrimination},
  author={Wang, Xudong and Liu, Ziwei and Yu, Stella X},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={12586--12595},
  year={2021}
}

License

This project is licensed under the MIT License. See LICENSE for more details. The parts described below follow their original license.

Acknowledgements

Part of this code is based on NPID, MoCo, CMC, infoMin and OpenSelfSup.