Network In Network Network In Network
Paper summary This paper studies a very natural generalization of convolutional layers by replacing a single filter that slides over the input feature map with a "micro network" (multi-layer perceptron). The authors argue that good abstractions are highly non-linear functions of input data and instead of generating an overcomplete number of feature maps and shrinking them down in higher layers (as is the case in traditional CNNs), it would be beneficial to generate better representations on each local patch, before feeding into the next layer. Main contributions: - Replaces the convolutional filter with a multi-layer perceptron. - Instead of fully connected layers, uses global average pooling. ## Strengths - Natural generalization of convolutional layers and thorough analysis. - Global average pooling of feature layers is easier to interpret and less prone to overfitting. - Better or at par with state-of-the-art classification results on CIFAR-10, CIFAR-100, SVHN, MNIST. ## Weaknesses / Notes - Should have explored NIN without dropout. - Results on ImageNet missing. - The global average pooling idea, although interpretable, doesn't seem to give easily to fine-tuning the network to other datasets. In finetuning, we usually replace and learn just the last layer.
Network In Network
Min Lin and Qiang Chen and Shuicheng Yan
arXiv e-Print archive - 2013 via Local arXiv
Keywords: cs.NE, cs.CV, cs.LG


Summary by Martin Thoma 4 years ago
Your comment: allows researchers to publish paper summaries that are voted on and ranked!

Sponsored by: and