Hi, I would like to make sure some implemental details on TinyImageNet-ResNet32 to CIFAR100-VGG9 section including:
The architecture of ResNet-32 (how to modify stem conv; others could be found in Figure 3)?
The architecture of VGG-9 (how to modify the classifier, use BN)?
Any change of the other hyperparameters, the learning rate is just fixed to 0.1 even though the batch size is 2x increased, an input image is smaller, and the capacity of networks is decreased?
How to set aside validation splits for CIFAR-100 and STL-10?
Please specify the augmentation strategy for 32x32 regimes.
We used the conventional residual networks used in the original ResNet paper.
We used the same architecture as the Jacobian Matching paper. (we used BN)
We used the same hyperparameters as described in the appendix in our paper. You should modify --pairs argument because VGG9 has a different number of layers.
In both cases, we sampled 10% validation samples from the training datasets.
We used the conventional augmentation scheme (random padding and flip.)
Hi, I would like to make sure some implemental details on
TinyImageNet-ResNet32
toCIFAR100-VGG9
section including:0.1
even though the batch size is 2x increased, an input image is smaller, and the capacity of networks is decreased?