Densely Connected Convolutional Networks (DenseNet)
Gao Huang, Zhuang Liu, Laurens van der Maaten, Kilian Q. Weinberger, CVPR-2017
Summary
This paper proposes a Densely Connected Convolutional Networks (DenseNet) which is an extension of the ResNet idea, in which every layer is connected to every other layer(hence the word Dense), and in which the combination is done via concatenation instead of simple summation.
It has significantly fewer parameters than ResNet and also achieved SOTA performance in many object recognition tasks.
This paper also got the Best Paper Award in CVPR 2017
Overview
Implementation Details
- They used three Denseblocks that had an equal number of layers. 3x3 convs with zero padding of 1 was used to preserve feature map size. 1x1 convs followed with 2x2 average pooling for transition layers .
- The feature-map sizes in the three dense blocks are 32×32, 16×16, and 8×8, respectively.
- Trained with SGD using Nesterov Momentum without dampening along with dropout.
Strengths
- As every layer is connected there is a lot of feature reuse and hence a small number of feature maps per layer (small k) also leads to good performance which is a clear improvement over ResNets. So it has fewer parameters and comparable accuracy.
- As a layer is closely connected with the output and input layers, so they kind of receive direct supervision and hence it improves gradient and information flow.
- They are also shown to have a regularizing nature and are less prone to over-fitting.
Weaknesses
- The regularisation nature of the network could have been explored better.
- The proper justification of using concatenation over summation of different activations could be explored further with more comprehensive set of experiments.
Implementation