SlideShare a Scribd company logo
Deep Learning
Presented By:
Srishty Saha
IIIT-DELHI
Shallow Learning
• SVM
• Linear & Kernel Regression
• Hidden Markov Models (HMM)
• Gaussian Mixture Models (GMM)
• Single hidden layer MLP
Limitations
Cannot make use of unlabeled data
Supervised vs Unsupervised Learning
• Supervised Learning
1.Output has to be produced according to target vector.
2.Input + Target vector = Training Pair
3.Labelled Data
• Unsupervised Learning ( self Organising)
1.Network receives input patterns to form clusters.
2.When a new input pattern is applied , output gives the class the input pattern
belongs to
3.Unlabelled Data
Neural Networks
• Machine Learning
• Knowledge from high dimensional data
• Classification
• Input: features of data
• supervised vs unsupervised
• labeled data
• Neurons
What is it used for?
• Classification
• Regression
---- Prediction
---- Fitting Curve
Multi Layer Perceptron
• Multiple Layers
• Feed Forward
• Connected Weights
• 1-of-N Output
hidden
output
Back Propagation
• Minimize error of
calculated output
• Adjust weights
• Gradient Descent
• Procedure
• Forward Phase
• Backpropagation
of errors
• For each sample,
multiple epochs
Problems with Backpropagation
• Multiple hidden Layers
• Get stuck in local optima
• start weights from random positions
• Only use labeled data
• most data is unlabeled
Deep Learning Means Feature Learning
• Deep Learning is about Learning Hierarchical Features.
Convolutional Neural Network
Feature extraction layer
Convolution layer
Shift and distortion invariance or
Subsampling layer
CNN contd.
• Detect the same feature at different positions in the
input image in C Layer.
features
CNN Contd.
Shared weights: all neurons in a feature share the
same weights (but not the biases).
In this way all neurons detect the same feature at
different positions in the input image.
Reduce the number of free parameters.
If a neuron in the feature map fires, this corresponds to a match with
the template
CNN Contd.
S Layer
The subsampling layers reduce the spatial resolution of each feature
map
By reducing the spatial resolution of the feature map, a certain degree
of shift and distortion invariance is achieved
Contd. S layer
The weight sharing is also applied in subsampling layers
Reduce the effect of noises and shift or distortion.
Applications
Speech Recognition.
Object Detection ( Computer Vision).
Web search – Text Analysis.
Few Insights Gathered From Papers.
• Used CBIR method to do feature extractions in Convolutional Layer.
• Applied filters to feature extraction.
• Used Definite size of patch to work upon.
• CNN method was used throughout.
• 3D convolution – time added as third factor .
• Feature Extraction so far observed was:
1. Gradient Filter in X and Y directions.
Object Detection
Architecture:
Dataset : MIT Face dataset 1104 faces.
Training – 200 images.
Test - 200 images.
The Convolutional Neural Network consists of two parts
1) the convolution layers and max-pooling layers
2) the fully connection layers and the output layers.
The Input Layer consists of 72x72 size histogram equalized images an
output is the set of different face images each of size 18x18.
The networks used for face detection and face recognition contains two
convolutional layer and two sub-sampling layer.
Output Layer
72 x 72 72x72 :5no 36x36: 5no
36X36
18 x 18:
12no
Output Layer
20 faces
Feature Maps :5no Feature Maps :5no
Input Layer Conv Layer-1
Kernels of 3x3
Conv Layer-2Samp Layer-1 Samp Layer-2
Convolutional Layer
Total of 5 kernels of size 3x3 is used to convolutional operation.
5 different feature maps :
• gray
• gradient –x
• gradient-y
• last two kernels gives the information below the eyes area.
Sampling Layer
• Mean filter of 2x2 is applied on image
• Alternate rows and alternate columns of image is sampled out.
72 x 72 72x72 :5no 36x36: 5no
36X36
18 x 18:
12no
Fully
connected
layer20 faces
Feature Maps :5no Feature Maps :5no
Input Layer Conv Layer-1
Kernels of 3x3
Conv Layer-2Samp Layer-1 Samp Layer-2
Fully Connected And Output Layer
• Output layer : 70 images.
72 x 72 72x72 :5no 36x36: 5no
36X36
18 x 18:
12no
Fully
connected
layer20 faces
Feature Maps :5no Feature Maps :5no
Input Layer Conv Layer-1
Kernels of 3x3
Conv Layer-2Samp Layer-1 Samp Layer-2
Error Propagation.
• Error Matrix e is obtained by finding difference between values of
neurons in output layer and fully connected layer.
• As there are 5 kernels in convolutional layers so each face will have 5
different feature maps. So in fully connected layer of Object
Recognition CNN,total neurons are 18 X 18 X 5(feature maps).
• So, Mean error M(i=1:5) of each Map is calculated.
• EM is Mean of M(i=1:5) is calculated.
• Error {M( i=1 :5) } is used for back-propagation.
• {EM} is used as a threshold such that any value below the threshold
is considered success or face match.
Implementation
1) Input of 200 images each of size 72X72 is presented to network one
by one for training.
2) In First Layer, Convolutional operation is performed using
aforementioned kernels of size 3X3 .The resultant output is of size
7 2 X 72 X 5.
3) In first S Layer, Sampling of image using mean filter of size 2X2
and sampling alternate rows and columns.The output of this of size
36 X 36 X 5.
4) In Second C layer,after convolution operation we get output
of size 36 X 36 X 5 .
5) In Second S layer, after sampling operation we get image of size
200 X 18 X 18 X 5.
6) The Fully connected layer is obtained after S layer and it is of
size 18 X 18 X 5 and each of neuron is connected to output layer.
7) Error Propagated using above method mentioned in
Back Propagation section of Object Recognition.
8) Error propagation takes place for fixed number of epochs in during
trainning.
9) For testing, {EM} obtained in equation is used as threshold to find face
match.
Results – Error vs Epochs
Results – Accuracy vs epochs
Object Detection.
• Input – Face images of size 72 x 72
Non Face images of size 72 x 72
• Output – 1 or 0
Error Propagation
• Error Matrix e is obtained by finding difference between values of
neurons in output layer and fully connected layer.
• Error of each neuron is propagated backwards and thus weight up-
dation is done.
• The backpropagation comes to hault when error <0.0003 or number
of epochs is 64 for training.
Object Detection ( Face/Non face)
1) Input of 50 (30 faces +20 non face) images each of size 72X72 is
presented to network one by one for training.
2) In First Layer, Convolutional operation is performed using
aforementioned kernels of size 3X3 .The resultant output is of size
7 2 X 72 X 5.
3) In first S Layer, Sampling of image using mean filter of size 2X2
and sampling alternate rows and columns.The output of this of size
36 X 36 X 5.
4) In Second C layer,after convolution operation we get output
of size 36 X 36 X 5 .
5) In Second S layer, after sampling operation we get image of size
18 X 18 X 5.
6) The Fully connected layer is obtained after S layer and it is of
size 18 X 18 X 5 and each of neuron is connected to output layer.
7) Error Propagated using above method mentioned in
Back Propagation section of Object Recognition.
8) Error propagation takes place for fixed number of epochs in during
trainning.
9) For testing, 200 images were used 125 faces and 75 non face.
Result:
Confusion Matrix Face Non face
Face (test) 100 25
Non Face (test) 32 43
Accuracy : 80% approx. for faces.
Accuracy : 57.33 % approx. for non faces.
Time to detect a face : 25.35 secs approx.
Object Recognition in an image.
Implementation iterative search
1) Input is an image is presented to network .
2) 72 X 72 patch is created and presented to network.
3) In First Layer, Convolutional operation is performed using
aforementioned kernels of size 3X3 .The resultant output is of size
7 2 X 72 X 5.
4) In first S Layer, Sampling of image using mean filter of size 2X2
and sampling alternate rows and columns.The output of this of size
36 X 36 X 5.
4) In Second C layer,after convolution operation we get output
of size 36 X 36 X 5 .
5) In Second S layer, after sampling operation we get image of size
200 X 18 X 18 X 5.
6) The Fully connected layer is obtained after S layer and it is of
size 18 X 18 X 5 and each of neuron is connected to output layer.
7) Error Propagated using above method mentioned in
Back Propagation section of Object Recognition.
8) Error propagation takes place for fixed number of epochs in during
trainning.
9) As an image is found Count = count+1.
Results: Error vs ith iteration face found.
Result : Accuracy
To do.
• Face detection using in set of images.
• Improve the accuracy.
• Implement it for videos.

More Related Content

PDF
Lecture 7: Recurrent Neural Networks
PPTX
Hardware Implementation of Cascade SVM
PDF
MapReduce based SVM
PPTX
Seminar_New -CESG
PDF
Svm map reduce_slides
PDF
Deep Neural Networks (D1L2 Insight@DCU Machine Learning Workshop 2017)
PDF
FPGA2018: A Lightweight YOLOv2: A binarized CNN with a parallel support vecto...
PPTX
A temporal classifier system using spiking neural networks
Lecture 7: Recurrent Neural Networks
Hardware Implementation of Cascade SVM
MapReduce based SVM
Seminar_New -CESG
Svm map reduce_slides
Deep Neural Networks (D1L2 Insight@DCU Machine Learning Workshop 2017)
FPGA2018: A Lightweight YOLOv2: A binarized CNN with a parallel support vecto...
A temporal classifier system using spiking neural networks

What's hot (20)

PDF
Gh2411361141
PPTX
Electricity price forecasting with Recurrent Neural Networks
PDF
ISCAS'18: A Deep Neural Network on the Nested RNS (NRNS) on an FPGA: Applied ...
PPTX
Deep Learning - RNN and CNN
PDF
Optimized Network-coded Scalable Video Multicasting over eMBMS Networks
PPTX
Bio-inspired Algorithms for Evolving the Architecture of Convolutional Neural...
PDF
PDF
Slide tesi
PDF
ISMVL2018: A Ternary Weight Binary Input Convolutional Neural Network
PDF
Optimizing Parallel Reduction in CUDA : NOTES
PDF
Batch normalization
PPTX
Introduction to Neural Network
PDF
Video Analysis with Recurrent Neural Networks (Master Computer Vision Barcelo...
PDF
Why Batch Normalization Works so Well
PDF
Pixel RNN to Pixel CNN++
PDF
The Perceptron (D1L1 Insight@DCU Machine Learning Workshop 2017)
PDF
FPL15 talk: Deep Convolutional Neural Network on FPGA
PPTX
Back propagation method
PDF
ujava.org Deep Learning with Convolutional Neural Network
PPTX
Convolutional Neural Networks
Gh2411361141
Electricity price forecasting with Recurrent Neural Networks
ISCAS'18: A Deep Neural Network on the Nested RNS (NRNS) on an FPGA: Applied ...
Deep Learning - RNN and CNN
Optimized Network-coded Scalable Video Multicasting over eMBMS Networks
Bio-inspired Algorithms for Evolving the Architecture of Convolutional Neural...
Slide tesi
ISMVL2018: A Ternary Weight Binary Input Convolutional Neural Network
Optimizing Parallel Reduction in CUDA : NOTES
Batch normalization
Introduction to Neural Network
Video Analysis with Recurrent Neural Networks (Master Computer Vision Barcelo...
Why Batch Normalization Works so Well
Pixel RNN to Pixel CNN++
The Perceptron (D1L1 Insight@DCU Machine Learning Workshop 2017)
FPL15 talk: Deep Convolutional Neural Network on FPGA
Back propagation method
ujava.org Deep Learning with Convolutional Neural Network
Convolutional Neural Networks
Ad

Similar to Deep learning and its application (20)

PPTX
Neural networks
PPTX
presentation of IntroductionDeepLearning.pptx
PDF
Hand gestures recognition seminar_ppt.pptx.pdf
PPT
Designed by Identity MLP
PPTX
Image Classification using deep learning
PPTX
Machine Learning DR PRKRao-PPT UNIT-II.pptx
PDF
Deep learning
PDF
dfdshofdifhdifhdfhgfoighfgofgfgfgfgdfdfdfdf
PPTX
Waste Classification System using Convolutional Neural Networks.pptx
PDF
IRJET - Hand Gesture Recognition to Perform System Operations
PPT
lec6a.ppt
PPTX
Facenet a unified embedding for fae recognition
PDF
Skin Lesion Detection from Dermoscopic Images using Convolutional Neural Netw...
PPTX
Back propagation network
PPTX
Trackster Pruning at the CMS High-Granularity Calorimeter
PPTX
Coin recognition using matlab
PDF
Improving face recognition by artificial neural network using principal compo...
PDF
Introduction to deeplearning engineering
PDF
Deep learning introduction basic information
PPTX
RNN-LSTM.pptx
Neural networks
presentation of IntroductionDeepLearning.pptx
Hand gestures recognition seminar_ppt.pptx.pdf
Designed by Identity MLP
Image Classification using deep learning
Machine Learning DR PRKRao-PPT UNIT-II.pptx
Deep learning
dfdshofdifhdifhdfhgfoighfgofgfgfgfgdfdfdfdf
Waste Classification System using Convolutional Neural Networks.pptx
IRJET - Hand Gesture Recognition to Perform System Operations
lec6a.ppt
Facenet a unified embedding for fae recognition
Skin Lesion Detection from Dermoscopic Images using Convolutional Neural Netw...
Back propagation network
Trackster Pruning at the CMS High-Granularity Calorimeter
Coin recognition using matlab
Improving face recognition by artificial neural network using principal compo...
Introduction to deeplearning engineering
Deep learning introduction basic information
RNN-LSTM.pptx
Ad

Recently uploaded (20)

PDF
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PDF
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
PDF
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
PDF
composite construction of structures.pdf
PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PPTX
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PDF
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
PPTX
CH1 Production IntroductoryConcepts.pptx
PPTX
UNIT-1 - COAL BASED THERMAL POWER PLANTS
PDF
Model Code of Practice - Construction Work - 21102022 .pdf
PPT
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PDF
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
PPTX
bas. eng. economics group 4 presentation 1.pptx
PPTX
Construction Project Organization Group 2.pptx
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
Evaluating the Democratization of the Turkish Armed Forces from a Normative P...
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Embodied AI: Ushering in the Next Era of Intelligent Systems
SM_6th-Sem__Cse_Internet-of-Things.pdf IOT
TFEC-4-2020-Design-Guide-for-Timber-Roof-Trusses.pdf
composite construction of structures.pdf
Automation-in-Manufacturing-Chapter-Introduction.pdf
MCN 401 KTU-2019-PPE KITS-MODULE 2.pptx
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
keyrequirementskkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkkk
CH1 Production IntroductoryConcepts.pptx
UNIT-1 - COAL BASED THERMAL POWER PLANTS
Model Code of Practice - Construction Work - 21102022 .pdf
CRASH COURSE IN ALTERNATIVE PLUMBING CLASS
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PRIZ Academy - 9 Windows Thinking Where to Invest Today to Win Tomorrow.pdf
bas. eng. economics group 4 presentation 1.pptx
Construction Project Organization Group 2.pptx
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks

Deep learning and its application

  • 2. Shallow Learning • SVM • Linear & Kernel Regression • Hidden Markov Models (HMM) • Gaussian Mixture Models (GMM) • Single hidden layer MLP Limitations Cannot make use of unlabeled data
  • 3. Supervised vs Unsupervised Learning • Supervised Learning 1.Output has to be produced according to target vector. 2.Input + Target vector = Training Pair 3.Labelled Data • Unsupervised Learning ( self Organising) 1.Network receives input patterns to form clusters. 2.When a new input pattern is applied , output gives the class the input pattern belongs to 3.Unlabelled Data
  • 4. Neural Networks • Machine Learning • Knowledge from high dimensional data • Classification • Input: features of data • supervised vs unsupervised • labeled data • Neurons
  • 5. What is it used for? • Classification • Regression ---- Prediction ---- Fitting Curve
  • 6. Multi Layer Perceptron • Multiple Layers • Feed Forward • Connected Weights • 1-of-N Output hidden output
  • 7. Back Propagation • Minimize error of calculated output • Adjust weights • Gradient Descent • Procedure • Forward Phase • Backpropagation of errors • For each sample, multiple epochs
  • 8. Problems with Backpropagation • Multiple hidden Layers • Get stuck in local optima • start weights from random positions • Only use labeled data • most data is unlabeled
  • 9. Deep Learning Means Feature Learning • Deep Learning is about Learning Hierarchical Features.
  • 10. Convolutional Neural Network Feature extraction layer Convolution layer Shift and distortion invariance or Subsampling layer
  • 11. CNN contd. • Detect the same feature at different positions in the input image in C Layer. features
  • 12. CNN Contd. Shared weights: all neurons in a feature share the same weights (but not the biases). In this way all neurons detect the same feature at different positions in the input image. Reduce the number of free parameters. If a neuron in the feature map fires, this corresponds to a match with the template
  • 13. CNN Contd. S Layer The subsampling layers reduce the spatial resolution of each feature map By reducing the spatial resolution of the feature map, a certain degree of shift and distortion invariance is achieved
  • 14. Contd. S layer The weight sharing is also applied in subsampling layers Reduce the effect of noises and shift or distortion.
  • 15. Applications Speech Recognition. Object Detection ( Computer Vision). Web search – Text Analysis.
  • 16. Few Insights Gathered From Papers. • Used CBIR method to do feature extractions in Convolutional Layer. • Applied filters to feature extraction. • Used Definite size of patch to work upon. • CNN method was used throughout. • 3D convolution – time added as third factor . • Feature Extraction so far observed was: 1. Gradient Filter in X and Y directions.
  • 17. Object Detection Architecture: Dataset : MIT Face dataset 1104 faces. Training – 200 images. Test - 200 images. The Convolutional Neural Network consists of two parts 1) the convolution layers and max-pooling layers 2) the fully connection layers and the output layers. The Input Layer consists of 72x72 size histogram equalized images an output is the set of different face images each of size 18x18. The networks used for face detection and face recognition contains two convolutional layer and two sub-sampling layer.
  • 18. Output Layer 72 x 72 72x72 :5no 36x36: 5no 36X36 18 x 18: 12no Output Layer 20 faces Feature Maps :5no Feature Maps :5no Input Layer Conv Layer-1 Kernels of 3x3 Conv Layer-2Samp Layer-1 Samp Layer-2
  • 19. Convolutional Layer Total of 5 kernels of size 3x3 is used to convolutional operation. 5 different feature maps : • gray • gradient –x • gradient-y • last two kernels gives the information below the eyes area.
  • 20. Sampling Layer • Mean filter of 2x2 is applied on image • Alternate rows and alternate columns of image is sampled out. 72 x 72 72x72 :5no 36x36: 5no 36X36 18 x 18: 12no Fully connected layer20 faces Feature Maps :5no Feature Maps :5no Input Layer Conv Layer-1 Kernels of 3x3 Conv Layer-2Samp Layer-1 Samp Layer-2
  • 21. Fully Connected And Output Layer • Output layer : 70 images. 72 x 72 72x72 :5no 36x36: 5no 36X36 18 x 18: 12no Fully connected layer20 faces Feature Maps :5no Feature Maps :5no Input Layer Conv Layer-1 Kernels of 3x3 Conv Layer-2Samp Layer-1 Samp Layer-2
  • 22. Error Propagation. • Error Matrix e is obtained by finding difference between values of neurons in output layer and fully connected layer. • As there are 5 kernels in convolutional layers so each face will have 5 different feature maps. So in fully connected layer of Object Recognition CNN,total neurons are 18 X 18 X 5(feature maps). • So, Mean error M(i=1:5) of each Map is calculated. • EM is Mean of M(i=1:5) is calculated. • Error {M( i=1 :5) } is used for back-propagation. • {EM} is used as a threshold such that any value below the threshold is considered success or face match.
  • 23. Implementation 1) Input of 200 images each of size 72X72 is presented to network one by one for training. 2) In First Layer, Convolutional operation is performed using aforementioned kernels of size 3X3 .The resultant output is of size 7 2 X 72 X 5. 3) In first S Layer, Sampling of image using mean filter of size 2X2 and sampling alternate rows and columns.The output of this of size 36 X 36 X 5.
  • 24. 4) In Second C layer,after convolution operation we get output of size 36 X 36 X 5 . 5) In Second S layer, after sampling operation we get image of size 200 X 18 X 18 X 5. 6) The Fully connected layer is obtained after S layer and it is of size 18 X 18 X 5 and each of neuron is connected to output layer. 7) Error Propagated using above method mentioned in Back Propagation section of Object Recognition. 8) Error propagation takes place for fixed number of epochs in during trainning. 9) For testing, {EM} obtained in equation is used as threshold to find face match.
  • 25. Results – Error vs Epochs
  • 26. Results – Accuracy vs epochs
  • 27. Object Detection. • Input – Face images of size 72 x 72 Non Face images of size 72 x 72 • Output – 1 or 0
  • 28. Error Propagation • Error Matrix e is obtained by finding difference between values of neurons in output layer and fully connected layer. • Error of each neuron is propagated backwards and thus weight up- dation is done. • The backpropagation comes to hault when error <0.0003 or number of epochs is 64 for training.
  • 29. Object Detection ( Face/Non face) 1) Input of 50 (30 faces +20 non face) images each of size 72X72 is presented to network one by one for training. 2) In First Layer, Convolutional operation is performed using aforementioned kernels of size 3X3 .The resultant output is of size 7 2 X 72 X 5. 3) In first S Layer, Sampling of image using mean filter of size 2X2 and sampling alternate rows and columns.The output of this of size 36 X 36 X 5.
  • 30. 4) In Second C layer,after convolution operation we get output of size 36 X 36 X 5 . 5) In Second S layer, after sampling operation we get image of size 18 X 18 X 5. 6) The Fully connected layer is obtained after S layer and it is of size 18 X 18 X 5 and each of neuron is connected to output layer. 7) Error Propagated using above method mentioned in Back Propagation section of Object Recognition. 8) Error propagation takes place for fixed number of epochs in during trainning. 9) For testing, 200 images were used 125 faces and 75 non face.
  • 31. Result: Confusion Matrix Face Non face Face (test) 100 25 Non Face (test) 32 43 Accuracy : 80% approx. for faces. Accuracy : 57.33 % approx. for non faces. Time to detect a face : 25.35 secs approx.
  • 33. Implementation iterative search 1) Input is an image is presented to network . 2) 72 X 72 patch is created and presented to network. 3) In First Layer, Convolutional operation is performed using aforementioned kernels of size 3X3 .The resultant output is of size 7 2 X 72 X 5. 4) In first S Layer, Sampling of image using mean filter of size 2X2 and sampling alternate rows and columns.The output of this of size 36 X 36 X 5.
  • 34. 4) In Second C layer,after convolution operation we get output of size 36 X 36 X 5 . 5) In Second S layer, after sampling operation we get image of size 200 X 18 X 18 X 5. 6) The Fully connected layer is obtained after S layer and it is of size 18 X 18 X 5 and each of neuron is connected to output layer. 7) Error Propagated using above method mentioned in Back Propagation section of Object Recognition. 8) Error propagation takes place for fixed number of epochs in during trainning. 9) As an image is found Count = count+1.
  • 35. Results: Error vs ith iteration face found.
  • 37. To do. • Face detection using in set of images. • Improve the accuracy. • Implement it for videos.