Concepedia

TLDR

Recent work has shown that convolutional networks can be substantially deeper, more accurate, and efficient to train if they contain shorter connections between layers close to the input and those close to the output. This paper introduces the Dense Convolutional Network (DenseNet), which connects each layer to every other layer in a feed‑forward fashion. DenseNet builds L(L+1)/2 direct connections by feeding each layer with all preceding feature maps and forwarding its own maps to all subsequent layers, and is evaluated on CIFAR‑10, CIFAR‑100, SVHN, and ImageNet. DenseNets alleviate vanishing gradients, strengthen feature propagation, encourage reuse, reduce parameters, and achieve state‑of‑the‑art performance on most benchmarks while using less memory and computation, with code and pre‑trained models available online.

Abstract

Recent work has shown that convolutional networks can be substantially deeper, more accurate, and efficient to train if they contain shorter connections between layers close to the input and those close to the output. In this paper, we embrace this observation and introduce the Dense Convolutional Network (DenseNet), which connects each layer to every other layer in a feed-forward fashion. Whereas traditional convolutional networks with L layers have L connections-one between each layer and its subsequent layer-our network has L(L+1)/2 direct connections. For each layer, the feature-maps of all preceding layers are used as inputs, and its own feature-maps are used as inputs into all subsequent layers. DenseNets have several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters. We evaluate our proposed architecture on four highly competitive object recognition benchmark tasks (CIFAR-10, CIFAR-100, SVHN, and ImageNet). DenseNets obtain significant improvements over the state-of-the-art on most of them, whilst requiring less memory and computation to achieve high performance. Code and pre-trained models are available at https://github.com/liuzhuang13/DenseNet.

References

YearCitations

2016

214.9K

2009

60.2K

1998

56.5K

2015

39.5K

2014

34.2K

2015

18.4K

2024

15.6K

1989

11.6K

2024

4.5K

2013

3.5K

Page 1