Closed NorbertZheng closed 1 year ago
Striving for Simplicity: The All Convolutional Net. All-CNN, by University of Freiburg. 2015 ICLR Workshop, Over 3800 Citations.
Convolutional Neural Network, CNN, Image Classification.
Convolution With Stride Larger Than 1 to Reduce Spatial Size More Aggressively (Figure from here).
There are two means suggested to replace pooling for spatial dimensionality reduction (i.e. along width & height axis, instead of filter axis):
The second option results in an increase of overall network parameters, yet without loss in accuracy.
The three base networks used for classification on CIFAR-10 and CIFAR-100.
Overall, three base models are suggested which consist only of convolutional layers with rectified linear non-linearities and an averaging + softmax layer to produce predictions over the whole image.
Model description of the three networks derived from base model C used for evaluating the importance of pooling in case of classification on CIFAR-10 and CIFAR-100.
Further enhanced models are derived from base models. The derived models for base models A and B are built analogously but not shown in the above table.
5×5 convolutions are replaced by 2 consecutive 3×3 convolutions.
Architecture of the Large All-CNN network for CIFAR-10.
The above shows the detailed architecture for CIFAR-10.
Architecture of the ImageNet network.
The above shows the detailed architecture for ImageNet.
All-CNN-C has the best performance.
Test error on CIFAR-10 and CIFAR-100 for the All-CNN compared to the state of the art from the literature.
On CIFAR-10, All-CNN is the All-CNN-C. It outperforms Maxout, and NIN, etc. On CIFAR-100, All-CNN-C obtains competitive performance.
An upscaled version of the All-CNN-B network is trained, which has 12 convolutional layers.
This network achieves a Top-1 validation error of 41.2% on ILSVRC-2012, when only evaluating on the center 224×224 patch, — which is comparable to the 40.7% Top-1 error reported by AlexNet.
(There are also sections to visualize the feature map response using deconvolutions which are something similar to ZFNet, please feel free to read the paper.)
Sik-Ho Tang. Review — Striving for Simplicity: The All Convolutional Net.