Closed NorbertZheng closed 1 year ago
Inthis story, Maxout Network, by Université de Montréal, is briefly reviewed. Ian J. GoodFellow, the first author, is also the inventor of Generative Adversarial Network (GAN). And Yoshua Bengio, the last author, just got the Turing Award recently this year (2019), which is the “Nobel Prize of computing”. These two authors plus the second last author Aaron Courville, three authors together, has also published the book “Deep learning”, through publisher MIT Press, in 2016 . And in this paper, for maxout network, it is published in 2013 ICML with over 1500 citations.
An MLP containing two maxout units.
$$ z{ij}=x^{T}W{...ij}+b_{ij}, $$
$$ h{i}(x)=\max{j\in [1,k]}z_{ij}, $$
$$ g(v)=h{1}(v)-h{2}(v), $$
The philosophy behind is that:
Test error on permutation invariant MNIST.
Test error on MNIST.
Test error on CIFAR-10.
CIFAR-100.
Test error on CIFAR-100.
Test error on SVHN.
It is interesting to read through the paper to see how authors make use of neural network to achieve the propositions above. And there are also ablation studies of Maxout Network against other activation functions like Tanh or ReLU at the end of the paper.
Sik-Ho Tang. Review: Maxout Network (Image Classification).