SlideShare a Scribd company logo
4
Most read
16
Most read
21
Most read
Densely Connected Convolutional
Networks
Authors: Gao Huang*, Zhuang Liu*, Laurens van der Maaten,
Kilian Q. Weinberger
Presented by:
Arun Kumar (ENG18CS0047)
Ananth Desai (ENG18CS0034)
Evans Nikith Royan (ENG18CS0054)
CNN
Convolutional neural networks (CNNs) have
become the dominant machine learning
approach for visual object recognition.
CNN
As CNNs become increasingly deep, a new
research problem emerges: as information about
the input or gradient passes through many
layers, it can vanish and “washout” by the time
it reaches the end (or beginning) of the network.
This problem is addressed by Resnets and by
Highway networks.
The combination of features using summation
may impede the information flow in the network.
ResNet
DenseNet
Traditional CNN
Resnet, Stochastic Depth, Densenet
Applications of Resnet: image recognition, localization,
and object detection
Stochastic depth: Better training of Resnet by dropping
layers randomly during training.
There is a great amount of redundancy in deep
(residual) networks.
There are direct connections from any layer to all
subsequent layers in Densenet.
A traditional CNN with L layers contains L connections. A
densenet with L layers contains L*(L+1)/2 connections.
Formula for Output at Each Layer
Consider a single image x0 that is passed through a CNN.
The network comprises L layers, each of which implements
a non-linear transformation Hℓ(·), where ℓ indexes the layer.
Hℓ(·) can contain operations such as BN, ReLU, Pool and
Conv.
In traditional CNN,
In Resnet,
In Densenet,
Hℓ(·) is a Composite Function
DenseNets
DenseNet layers are very narrow (eg: 12 filters
per layer), adding only a small set of feature-
maps to the “collective knowledge” of the
network and keep the remaining feature-maps
unchanged; the final classifier makes a decision
based on all feature-maps in the network.
DenseNets exploit the potential of the network
through feature reuse.
Inception networks also concatenate features
from different layers.
9 / 21
Pooling Layers
The concatenation operation is not viable when the size
of feature-maps changes.
Pooling can be either max-pool or average-pool and
reduces the size of the feature-maps.
To facilitate down-sampling in the architecture the
network is divided into multiple densely connected
dense blocks with pooling layers inserted between them.
10 / 21
Growth Rate
● If each function Hℓ
produces k featuremaps, it follows
that the ℓth
layer has k0
+k ×(ℓ−1) input feature-maps,
where k0
is the number of channels in the input layer.
●
k is referred to as the growth rate of the network.
●
The growth rate regulates how much new information
each layer contributes to the global state. The global
state, once written, can be accessed from everywhere
within the network and, unlike in traditional network
architectures, there is no need to replicate it from layer
to layer.
11 / 21
Bottleneck layers
●
Although each layer only produces k output
feature-maps, it typically has many more inputs.
●
A 1×1 convolution can be introduced as bottleneck
layer before each 3×3 convolution to reduce the
number of input feature-maps, and thus to improve
computational efficiency.
12 / 21
Bottleneck layers
13 / 21
Compression
●
To further improve model compactness, the number
of feature-maps is reduced at transition layers.
●
If a dense block contains m feature-maps, we let
the following transition layer generate [θm] output
feature maps, where 0 < θ ≤1 is referred to as the
compression factor.
14 / 21
Implementation
●
The DenseNet used in the experiments has three
dense blocks that each has an equal number of
layers.
●
Before entering the first dense block, a convolution
with 16 output channels is performed on the input
images.
●
A 1×1 convolution followed by 2×2 average pooling
is used as transition layers between two
contiguous dense blocks.
●
At the end of the last dense block, a global average
pooling is performed and then a softmax classifier
is attached.
15 / 21
ADVANTAGES OF DENSE CONNECTIVITY
●
ADVANTAGE 1: STRONG GRADIENT FLOW
16 / 21
●
ADVANTAGE 2: PARAMETER & COMPUTATIONAL
EFFICIENCY
17 / 21
●
ADVANTAGE 3: MAINTAINS LOW COMPLEXITY
FEATURES
18 / 21
19 / 21
RESULTS
●
Results on CIFAR-10:
20 / 21
●
Results on CIFAR-100:
21 / 21
●
Results on IMAGENET

More Related Content

PDF
GoogLeNet Insights
PPTX
PPTX
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
PDF
Mobilenetv1 v2 slide
PPTX
Batch normalization presentation
PDF
PPTX
Squeeze Excitation Networks, The simple idea that won the final ImageNet Chal...
PDF
CNN Algorithm
GoogLeNet Insights
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
Mobilenetv1 v2 slide
Batch normalization presentation
Squeeze Excitation Networks, The simple idea that won the final ImageNet Chal...
CNN Algorithm

What's hot (20)

PDF
Densely Connected Convolutional Networks
PPTX
Convolutional Neural Network and Its Applications
PPTX
Convolutional Neural Network - CNN | How CNN Works | Deep Learning Course | S...
PPTX
PDF
[PR12] You Only Look Once (YOLO): Unified Real-Time Object Detection
PPTX
CONVOLUTIONAL NEURAL NETWORK
PPT
K means Clustering Algorithm
PDF
Deep learning based object detection basics
PPTX
CNN Tutorial
PDF
Autoencoders
PDF
Training Neural Networks
PDF
Generative adversarial networks
PPTX
Convolutional Neural Networks
PPTX
Generative Adversarial Network (GAN)
PDF
An introduction to Deep Learning
PPT
Cnn method
PPSX
Perceptron (neural network)
PPTX
Resnet.pptx
PPTX
Activation functions and Training Algorithms for Deep Neural network
PDF
Convolutional Neural Networks (CNN)
Densely Connected Convolutional Networks
Convolutional Neural Network and Its Applications
Convolutional Neural Network - CNN | How CNN Works | Deep Learning Course | S...
[PR12] You Only Look Once (YOLO): Unified Real-Time Object Detection
CONVOLUTIONAL NEURAL NETWORK
K means Clustering Algorithm
Deep learning based object detection basics
CNN Tutorial
Autoencoders
Training Neural Networks
Generative adversarial networks
Convolutional Neural Networks
Generative Adversarial Network (GAN)
An introduction to Deep Learning
Cnn method
Perceptron (neural network)
Resnet.pptx
Activation functions and Training Algorithms for Deep Neural network
Convolutional Neural Networks (CNN)
Ad

Similar to Densenet CNN (20)

PPTX
Introduction to CNN Models: DenseNet & MobileNet
PDF
Convolutional Neural Networks : Popular Architectures
PDF
Computer vision for transportation
PPTX
adlkchiuabcndjhvkajnfdkjhcfatgcbajkbcyudfctauygb
PPTX
Deep learning and computer vision
PDF
CONTRAST OF RESNET AND DENSENET BASED ON THE RECOGNITION OF SIMPLE FRUIT DATA...
PDF
6119ijcsitce01
PDF
International Journal of Computational Science, Information Technology and Co...
PDF
CONTRAST OF RESNET AND DENSENET BASED ON THE RECOGNITION OF SIMPLE FRUIT DATA...
PPTX
DL-CO2-Session6-VGGNet_GoogLeNet_ResNet_DenseNet_RCNN.pptx
PPTX
Tìm hiểu về CNN và ResNet | Computer Vision
PDF
PDF
DLD meetup 2017, Efficient Deep Learning
PDF
Architecture Design for Deep Neural Networks I
PPTX
FINAL_Team_4.pptx
PDF
Image Classification on ImageNet (D1L3 Insight@DCU Machine Learning Workshop ...
PDF
CNNs: from the Basics to Recent Advances
PDF
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
PPTX
CNN Arcitecture Implementation Resnet CNN-RESNET
PPTX
conv_nets.pptx
Introduction to CNN Models: DenseNet & MobileNet
Convolutional Neural Networks : Popular Architectures
Computer vision for transportation
adlkchiuabcndjhvkajnfdkjhcfatgcbajkbcyudfctauygb
Deep learning and computer vision
CONTRAST OF RESNET AND DENSENET BASED ON THE RECOGNITION OF SIMPLE FRUIT DATA...
6119ijcsitce01
International Journal of Computational Science, Information Technology and Co...
CONTRAST OF RESNET AND DENSENET BASED ON THE RECOGNITION OF SIMPLE FRUIT DATA...
DL-CO2-Session6-VGGNet_GoogLeNet_ResNet_DenseNet_RCNN.pptx
Tìm hiểu về CNN và ResNet | Computer Vision
DLD meetup 2017, Efficient Deep Learning
Architecture Design for Deep Neural Networks I
FINAL_Team_4.pptx
Image Classification on ImageNet (D1L3 Insight@DCU Machine Learning Workshop ...
CNNs: from the Basics to Recent Advances
AlexNet(ImageNet Classification with Deep Convolutional Neural Networks)
CNN Arcitecture Implementation Resnet CNN-RESNET
conv_nets.pptx
Ad

Recently uploaded (20)

PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PDF
Accuracy of neural networks in brain wave diagnosis of schizophrenia
PPTX
Spectroscopy.pptx food analysis technology
PDF
Getting Started with Data Integration: FME Form 101
PPTX
Programs and apps: productivity, graphics, security and other tools
PDF
Encapsulation theory and applications.pdf
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PDF
gpt5_lecture_notes_comprehensive_20250812015547.pdf
PPTX
Tartificialntelligence_presentation.pptx
PDF
A comparative analysis of optical character recognition models for extracting...
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
cuic standard and advanced reporting.pdf
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Network Security Unit 5.pdf for BCA BBA.
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
Advanced methodologies resolving dimensionality complications for autism neur...
Building Integrated photovoltaic BIPV_UPV.pdf
Accuracy of neural networks in brain wave diagnosis of schizophrenia
Spectroscopy.pptx food analysis technology
Getting Started with Data Integration: FME Form 101
Programs and apps: productivity, graphics, security and other tools
Encapsulation theory and applications.pdf
Dropbox Q2 2025 Financial Results & Investor Presentation
“AI and Expert System Decision Support & Business Intelligence Systems”
gpt5_lecture_notes_comprehensive_20250812015547.pdf
Tartificialntelligence_presentation.pptx
A comparative analysis of optical character recognition models for extracting...
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Unlocking AI with Model Context Protocol (MCP)
cuic standard and advanced reporting.pdf
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Diabetes mellitus diagnosis method based random forest with bat algorithm
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Network Security Unit 5.pdf for BCA BBA.
Profit Center Accounting in SAP S/4HANA, S4F28 Col11

Densenet CNN

  • 1. Densely Connected Convolutional Networks Authors: Gao Huang*, Zhuang Liu*, Laurens van der Maaten, Kilian Q. Weinberger Presented by: Arun Kumar (ENG18CS0047) Ananth Desai (ENG18CS0034) Evans Nikith Royan (ENG18CS0054)
  • 2. CNN Convolutional neural networks (CNNs) have become the dominant machine learning approach for visual object recognition.
  • 3. CNN As CNNs become increasingly deep, a new research problem emerges: as information about the input or gradient passes through many layers, it can vanish and “washout” by the time it reaches the end (or beginning) of the network. This problem is addressed by Resnets and by Highway networks. The combination of features using summation may impede the information flow in the network.
  • 5. Resnet, Stochastic Depth, Densenet Applications of Resnet: image recognition, localization, and object detection Stochastic depth: Better training of Resnet by dropping layers randomly during training. There is a great amount of redundancy in deep (residual) networks. There are direct connections from any layer to all subsequent layers in Densenet. A traditional CNN with L layers contains L connections. A densenet with L layers contains L*(L+1)/2 connections.
  • 6. Formula for Output at Each Layer Consider a single image x0 that is passed through a CNN. The network comprises L layers, each of which implements a non-linear transformation Hℓ(·), where ℓ indexes the layer. Hℓ(·) can contain operations such as BN, ReLU, Pool and Conv. In traditional CNN, In Resnet, In Densenet,
  • 7. Hℓ(·) is a Composite Function
  • 8. DenseNets DenseNet layers are very narrow (eg: 12 filters per layer), adding only a small set of feature- maps to the “collective knowledge” of the network and keep the remaining feature-maps unchanged; the final classifier makes a decision based on all feature-maps in the network. DenseNets exploit the potential of the network through feature reuse. Inception networks also concatenate features from different layers.
  • 9. 9 / 21 Pooling Layers The concatenation operation is not viable when the size of feature-maps changes. Pooling can be either max-pool or average-pool and reduces the size of the feature-maps. To facilitate down-sampling in the architecture the network is divided into multiple densely connected dense blocks with pooling layers inserted between them.
  • 10. 10 / 21 Growth Rate ● If each function Hℓ produces k featuremaps, it follows that the ℓth layer has k0 +k ×(ℓ−1) input feature-maps, where k0 is the number of channels in the input layer. ● k is referred to as the growth rate of the network. ● The growth rate regulates how much new information each layer contributes to the global state. The global state, once written, can be accessed from everywhere within the network and, unlike in traditional network architectures, there is no need to replicate it from layer to layer.
  • 11. 11 / 21 Bottleneck layers ● Although each layer only produces k output feature-maps, it typically has many more inputs. ● A 1×1 convolution can be introduced as bottleneck layer before each 3×3 convolution to reduce the number of input feature-maps, and thus to improve computational efficiency.
  • 13. 13 / 21 Compression ● To further improve model compactness, the number of feature-maps is reduced at transition layers. ● If a dense block contains m feature-maps, we let the following transition layer generate [θm] output feature maps, where 0 < θ ≤1 is referred to as the compression factor.
  • 14. 14 / 21 Implementation ● The DenseNet used in the experiments has three dense blocks that each has an equal number of layers. ● Before entering the first dense block, a convolution with 16 output channels is performed on the input images. ● A 1×1 convolution followed by 2×2 average pooling is used as transition layers between two contiguous dense blocks. ● At the end of the last dense block, a global average pooling is performed and then a softmax classifier is attached.
  • 15. 15 / 21 ADVANTAGES OF DENSE CONNECTIVITY ● ADVANTAGE 1: STRONG GRADIENT FLOW
  • 16. 16 / 21 ● ADVANTAGE 2: PARAMETER & COMPUTATIONAL EFFICIENCY
  • 17. 17 / 21 ● ADVANTAGE 3: MAINTAINS LOW COMPLEXITY FEATURES
  • 20. 20 / 21 ● Results on CIFAR-100:
  • 21. 21 / 21 ● Results on IMAGENET