htqin / IR-Net

[CVPR 2020] This project is the PyTorch implementation of our accepted CVPR 2020 paper : forward and backward information retention for accurate binary neural networks.
177 stars 38 forks source link

Details about ImageNet experiments setting #6

Closed licj15 closed 4 years ago

licj15 commented 4 years ago

@htqin Thank you for your quick response! I intended to ask this in this issue, but it was closed and this seems to be a new question, so I open a new issue for this question. Thank you a lot!

I notice in the yaml

    augmentation:
        input_size: 224
        test_resize: 256
        colorjitter: [0.2, 0.2, 0.2, 0.1] 
  1. Does it mean you use 224224 for train, and upsample it to 256256 during test? If so, I noticed you didn't mention so in your paper, do you have any references for this? I was wondering if such techniques are also used in baselines.
  2. Do you only use "colorjitter: [0.2, 0.2, 0.2, 0.1]" for augmentation? Do you use random crop or flip? Because I noticed you use such augmentation in CIFAR-10, not sure if you use any other augmentation beside "colorjitter: [0.2, 0.2, 0.2, 0.1]" for ImageNet.

Thank you! Your efforts and the sprit to open source will benefit the community a lot!

Haleski47 commented 4 years ago

@htqin Thank you for your quick response! I intended to ask this in this issue, but it was closed and this seems to be a new question, so I open a new issue for this question. Thank you a lot!

I notice in the yaml

    augmentation:
        input_size: 224
        test_resize: 256
        colorjitter: [0.2, 0.2, 0.2, 0.1] 
  1. Does it mean you use 224224 for train, and upsample it to 256256 during test? If so, I noticed you didn't mention so in your paper, do you have any references for this? I was wondering if such techniques are also used in baselines.
  2. Do you only use "colorjitter: [0.2, 0.2, 0.2, 0.1]" for augmentation? Do you use random crop or flip? Because I noticed you use such augmentation in CIFAR-10, not sure if you use any other augmentation beside "colorjitter: [0.2, 0.2, 0.2, 0.1]" for ImageNet.

Thank you! Your efforts and the sprit to open source will benefit the community a lot!

i'm also curious about the training details for imagenet dataset. i see the traing learning rate is 0.8 for warmup stage from the taml file. it's a bit crazy...

htqin commented 4 years ago

Thanks for your attention! @licj15 @Haleski47 A1: We use the same augmentation setting on both full-precision and binary networks. These settings are mainly for full-precision networks and seem to improve the performance of baselines (including augmentation and warmup lr). Thank you for pointing out, we will claim our various settings more clearly in future work. A2: [Q2: "if you use any other augmentation beside "colorjitter: [0.2, 0.2, 0.2, 0.1]" for ImageNet"] No, the networks for CIFAR-10 and imagenet are based on different projects, and we follow their setting of augmentation. (ResNet CIFAR-10: https://github.com/akamaster/pytorch_resnet_cifar10)