SlideShare a Scribd company logo
Use and distribution limited solely to authorized personnel. (c) Copyright 2018
Lecture Series: AI is the New Electricity?
AILABS Academy
J-3, GP Block, Sector V, Salt Lake City,
Kolkata, West Bengal 700091
Classification and Estimation using
Artificial Neural Network
Presented at AILABS Academy,
Kolkata on April 18th 2018
Dinabandhu Bhandari
Contents
Examples of Classification and Estimation
Components of Learning
ANN – Architecture and Back Propagation
Classification and Estimation
Some Industrial Problems
Observation
Discussion and Q&A
Paragraph Classification
An article without any style of the paragraphs
The article after applying appropriate style
Daily Roadway Development for Mining
Serial
No
No of
Agitator
Trucks
[1 to 4]
Agitator
Capacity m3)
[5 to 10 m3]
No of
Kibble
Trucks
[1 to 5]
Kibble
Capacity m3)
[2 to 5 m3]
Distance to
Site (m)
[1 to 4 km]
Daily Shotcrete
Roadway
Development
(m3)
1 2 5 2 5 1000 49.93
2 2 5 2 5 1500 49.93
8 1 5 2 2.5 1500 24.98
9 1 5 2 2.5 2000 24.98
10 1 5 2 2.5 2500 24.98
79 2 5 1 4 1000 27.95
80 2 5 1 4 1500 23.95
81 2 5 1 4 2000 23.31
82 2 5 1 4 2500 19.96
83 2 5 1 4 3000 19.31
89 1 5 1 4 3000 11.98
90 1 5 1 4 3500 11.98
Bank loan approval
Feature Value
Age 35 years
Gender Male
Debt 10,00,000
Salary 15,00,000
Years in home 5 years
… …
1. Approve or not?
2. How Much?
Classification and Estimation
Formulation – Classification System
Formulation – Learning (Classification)
System
Learning Paradigm
The linear model - Perceptron
The linear model - Perceptron
Perceptron Learning Algorithm
Perceptron Learning Algorithm
Artificial Neural Network (Feed-forward network)?
 A set of neurons (simple processing unit) arranged in number
of layers.
 The neurons in each layer feed their output forward to the next
layer until we get the final output from the neural network.
 There can be any number of hidden layers within a
Feedforward network.
 The number of neurons can be completely arbitrary.
16
Neuron – a simple processing unit
 A neuron can have any number of inputs
 The inputs may be represented as x1, x2, x3… xn.
 And the corresponding weights for the inputs as w1,
w2, w3… wn.
 Output a = x1w1+x2w2+x3w3... +xnwn
17
• No connections within a layer
• No direct connections between input and output layers
• Fully connected between layers
• Number of output units need not equal number of input units
• Number of hidden units per layer can be more or less than
input or output units
y f w x bi ij j i
j
m
= +
=
( )
1
Often include bias as an extra weight
18
Properties of architecture
Perceptron Learning Theorem
 Recap: A perceptron (threshold unit) can learn
anything that it can represent (i.e. anything separable
with a hyperplane)
19
The Exclusive OR problem
 A Perceptron cannot represent Exclusive OR since it is
not linearly separable.
20
21
Minsky & Papert (1969) offered solution to XOR problem by
combining perceptron unit responses using a second layer of
Units. Piecewise linear classification using an MLP with
threshold (perceptron) units
1
2
+1
+1
3
22
What do each of the layers do?
1st layer draws
linear boundaries
2nd layer combines
the boundaries
3rd layer can generate
arbitrarily complex
boundaries
23
2. Backward pass phase: computes ‘error signal’, propagates
the error backwards through network starting at output units
(where the error is the difference between actual and desired
output values)
1. Forward pass phase: computes ‘functional signal’, feed forward
propagation of input pattern signals through network
Two phases of Backpropagation learning algorithm ‘BP’
24
Frank Rosenblatt in 1961
Backpropagation
xk
xi
wki
wjk
dj
dk
yj
Backward step:
propagate errors from
output to hidden layer
Forward step:
Propagate activation
from input to output
layer
Backpropagation Algorithm
 Initialize each wi to some small random value
 Until the termination condition is met, Do
 For each training example <(x1,…xn),t> Do
 Input the instance (x1,…,xn) to the network and compute
the network outputs yk
 For each output unit k
 dk=yk(1-yk)(tk-yk)
 For each hidden unit h
 dh=yh(1-yh) k wh,k dk
 For each network weight wi,j Do
 wi,j=wi,j+wi,j where
wi,j=  dj xi,j
Backpropagation
 Gradient descent over entire network weight vector
 Easily generalized to arbitrary directed graphs
 Finds a local, not necessarily global error minimum
-in practice often works well (can be invoked multiple times
with different initial weights)
 Often include weight momentum term
wi,j(n)=  dj xi,j +  wi,j (n-1)
 Minimizes error for training examples
 Will it generalize well to unseen instances (out of sample
error; over-fitting)?
 Training can be slow typical 1000-10000 iterations
Convergence of Backprop
 Gradient descent converges to some local minimum perhaps
not global minimum
 Add momentum term: wki(n)
 wki(n) =  dk(n) xi (n) +  wki(n-1)
with   [0,1]
 Stochastic gradient descent
 Train multiple nets with different initial weights
Ways to use weight derivatives
 How often to update
 after each training case?
 after a full sweep through the training data?
 How much to update
 Use a fixed learning rate?
 Adapt the learning rate?
 Add momentum?
Training
 This was a single iteration of back-prop
 Training requires many iterations with many
training examples or epochs (one epoch is entire
presentation of complete training set)
 It can be slow !
 Parallel computation implementation is also
possible
30
Training and testing data
 How many examples ?
 The more the merrier !
 Disjoint training and testing data sets
 learn from training data but evaluate performance
(generalization ability) on unseen test data
 Aim: minimize error on test data
31
Hoeffding’s Inequality
In-sample error Out-of-
sample error
|H|, cardinility of
hypothesis set
Number of
training samples
Estimation
xk
xi
wki
wjk
dj
dk
yj
Difference with Classifier:
Only in the calculation
of error
References
Thank
You

More Related Content

PDF
Joint unsupervised learning of deep representations and image clusters
PPT
2.5 backpropagation
PDF
15 Machine Learning Multilayer Perceptron
PDF
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
PPTX
Back propagation method
PPT
Multi-Layer Perceptrons
PDF
Deep Learning for Computer Vision: Backward Propagation (UPC 2016)
PPTX
Multilayer perceptron
Joint unsupervised learning of deep representations and image clusters
2.5 backpropagation
15 Machine Learning Multilayer Perceptron
Optimizing Deep Networks (D1L6 Insight@DCU Machine Learning Workshop 2017)
Back propagation method
Multi-Layer Perceptrons
Deep Learning for Computer Vision: Backward Propagation (UPC 2016)
Multilayer perceptron

What's hot (20)

PDF
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
PDF
Deep Learning for Computer Vision: Deep Networks (UPC 2016)
PDF
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
PDF
Multi Layer Perceptron & Back Propagation
PDF
Artificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
PDF
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
PDF
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
PPTX
Neural net and back propagation
PDF
(Kpi summer school 2015) theano tutorial part2
PDF
Optimization for Deep Networks (D2L1 2017 UPC Deep Learning for Computer Vision)
PDF
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
PDF
Deep learning for molecules, introduction to chainer chemistry
PDF
Deep Learning for Computer Vision: Segmentation (UPC 2016)
DOCX
Backpropagation
PDF
Deep Neural Networks (D1L2 Insight@DCU Machine Learning Workshop 2017)
PDF
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
PDF
Convolutional Neural Networks (DLAI D5L1 2017 UPC Deep Learning for Artificia...
PDF
The Perceptron (D1L2 Deep Learning for Speech and Language)
PDF
Deep Learning for Computer Vision: Attention Models (UPC 2016)
PPTX
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Training Deep Networks with Backprop (D1L4 Insight@DCU Machine Learning Works...
Deep Learning for Computer Vision: Deep Networks (UPC 2016)
Deep Learning for Computer Vision: Data Augmentation (UPC 2016)
Multi Layer Perceptron & Back Propagation
Artificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
Unsupervised Deep Learning (D2L1 Insight@DCU Machine Learning Workshop 2017)
Generative Models and Adversarial Training (D2L3 Insight@DCU Machine Learning...
Neural net and back propagation
(Kpi summer school 2015) theano tutorial part2
Optimization for Deep Networks (D2L1 2017 UPC Deep Learning for Computer Vision)
Convolutional Neural Networks - Veronica Vilaplana - UPC Barcelona 2018
Deep learning for molecules, introduction to chainer chemistry
Deep Learning for Computer Vision: Segmentation (UPC 2016)
Backpropagation
Deep Neural Networks (D1L2 Insight@DCU Machine Learning Workshop 2017)
Unsupervised Learning (D2L6 2017 UPC Deep Learning for Computer Vision)
Convolutional Neural Networks (DLAI D5L1 2017 UPC Deep Learning for Artificia...
The Perceptron (D1L2 Deep Learning for Speech and Language)
Deep Learning for Computer Vision: Attention Models (UPC 2016)
Multilayer Perceptron (DLAI D1L2 2017 UPC Deep Learning for Artificial Intell...
Ad

Similar to AILABS - Lecture Series - Is AI the New Electricity? Topic:- Classification and Estimation using Artificial Neural Network, Presented by - Prof. Dinabandhu Bhandari (20)

PPT
Artificial Neural Network
PPT
ann-ics320Part4.ppt
PPT
ann-ics320Part4.ppt
PPTX
linearly separable and therefore a set of weights exist that are consistent ...
PPTX
Introduction to Perceptron and Neural Network.pptx
PDF
Classification by back propagation, multi layered feed forward neural network...
PDF
lecture24NeuralNetworksNYUcourseEEEE.pdf
PDF
Deep Learning Study _ FInalwithCNN_RNN_LSTM_GRU.pdf
PPTX
Chapter-5-Part I-Basics-Neural-Networks.pptx
PPTX
Feed forward back propogation algorithm .pptx
PPTX
Artificial Neural Networks presentations
PPT
Data Mining: Concepts and techniques classification _chapter 9 :advanced methods
PPT
Neural
PPTX
Deeplearning for Computer Vision PPT with
PPT
Chapter 9. Classification Advanced Methods.ppt
PDF
09 classadvanced
PPTX
03 Single layer Perception Classifier
PPTX
10 Backpropagation Algorithm for Neural Networks (1).pptx
PPT
19_Learning.ppt
Artificial Neural Network
ann-ics320Part4.ppt
ann-ics320Part4.ppt
linearly separable and therefore a set of weights exist that are consistent ...
Introduction to Perceptron and Neural Network.pptx
Classification by back propagation, multi layered feed forward neural network...
lecture24NeuralNetworksNYUcourseEEEE.pdf
Deep Learning Study _ FInalwithCNN_RNN_LSTM_GRU.pdf
Chapter-5-Part I-Basics-Neural-Networks.pptx
Feed forward back propogation algorithm .pptx
Artificial Neural Networks presentations
Data Mining: Concepts and techniques classification _chapter 9 :advanced methods
Neural
Deeplearning for Computer Vision PPT with
Chapter 9. Classification Advanced Methods.ppt
09 classadvanced
03 Single layer Perception Classifier
10 Backpropagation Algorithm for Neural Networks (1).pptx
19_Learning.ppt
Ad

More from AILABS Academy (6)

PDF
AILABS - Lecture Series - Is AI the New Electricity? - Advances In Machine Le...
PDF
AILABS - Lecture Series - Is AI the New Electricity? Topic: Intelligent proce...
PDF
AILABS - Lecture Series - Is AI the New Electricity? Topic:- Interplay of Tr...
PDF
AILABS - Lecture Series - Is AI the New Electricity. Topic- Role of AI in Log...
PDF
AILABS Lecture Series - Is AI The New Electricity. Topic - Deep Learning - Ev...
PDF
Nano Degree On Deep Learning
AILABS - Lecture Series - Is AI the New Electricity? - Advances In Machine Le...
AILABS - Lecture Series - Is AI the New Electricity? Topic: Intelligent proce...
AILABS - Lecture Series - Is AI the New Electricity? Topic:- Interplay of Tr...
AILABS - Lecture Series - Is AI the New Electricity. Topic- Role of AI in Log...
AILABS Lecture Series - Is AI The New Electricity. Topic - Deep Learning - Ev...
Nano Degree On Deep Learning

Recently uploaded (20)

PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Review of recent advances in non-invasive hemoglobin estimation
PPTX
Big Data Technologies - Introduction.pptx
PDF
Unlocking AI with Model Context Protocol (MCP)
PPTX
Spectroscopy.pptx food analysis technology
PDF
cuic standard and advanced reporting.pdf
PPTX
Programs and apps: productivity, graphics, security and other tools
PDF
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
PPT
Teaching material agriculture food technology
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PPTX
MYSQL Presentation for SQL database connectivity
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
Electronic commerce courselecture one. Pdf
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PDF
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
DOCX
The AUB Centre for AI in Media Proposal.docx
PDF
KodekX | Application Modernization Development
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Review of recent advances in non-invasive hemoglobin estimation
Big Data Technologies - Introduction.pptx
Unlocking AI with Model Context Protocol (MCP)
Spectroscopy.pptx food analysis technology
cuic standard and advanced reporting.pdf
Programs and apps: productivity, graphics, security and other tools
7 ChatGPT Prompts to Help You Define Your Ideal Customer Profile.pdf
Teaching material agriculture food technology
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
MYSQL Presentation for SQL database connectivity
Building Integrated photovoltaic BIPV_UPV.pdf
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
Electronic commerce courselecture one. Pdf
Understanding_Digital_Forensics_Presentation.pptx
How UI/UX Design Impacts User Retention in Mobile Apps.pdf
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
The AUB Centre for AI in Media Proposal.docx
KodekX | Application Modernization Development
Advanced methodologies resolving dimensionality complications for autism neur...

AILABS - Lecture Series - Is AI the New Electricity? Topic:- Classification and Estimation using Artificial Neural Network, Presented by - Prof. Dinabandhu Bhandari

  • 1. Use and distribution limited solely to authorized personnel. (c) Copyright 2018 Lecture Series: AI is the New Electricity? AILABS Academy J-3, GP Block, Sector V, Salt Lake City, Kolkata, West Bengal 700091 Classification and Estimation using Artificial Neural Network Presented at AILABS Academy, Kolkata on April 18th 2018 Dinabandhu Bhandari
  • 2. Contents Examples of Classification and Estimation Components of Learning ANN – Architecture and Back Propagation Classification and Estimation Some Industrial Problems Observation Discussion and Q&A
  • 4. An article without any style of the paragraphs
  • 5. The article after applying appropriate style
  • 6. Daily Roadway Development for Mining Serial No No of Agitator Trucks [1 to 4] Agitator Capacity m3) [5 to 10 m3] No of Kibble Trucks [1 to 5] Kibble Capacity m3) [2 to 5 m3] Distance to Site (m) [1 to 4 km] Daily Shotcrete Roadway Development (m3) 1 2 5 2 5 1000 49.93 2 2 5 2 5 1500 49.93 8 1 5 2 2.5 1500 24.98 9 1 5 2 2.5 2000 24.98 10 1 5 2 2.5 2500 24.98 79 2 5 1 4 1000 27.95 80 2 5 1 4 1500 23.95 81 2 5 1 4 2000 23.31 82 2 5 1 4 2500 19.96 83 2 5 1 4 3000 19.31 89 1 5 1 4 3000 11.98 90 1 5 1 4 3500 11.98
  • 7. Bank loan approval Feature Value Age 35 years Gender Male Debt 10,00,000 Salary 15,00,000 Years in home 5 years … … 1. Approve or not? 2. How Much?
  • 10. Formulation – Learning (Classification) System
  • 12. The linear model - Perceptron
  • 13. The linear model - Perceptron
  • 16. Artificial Neural Network (Feed-forward network)?  A set of neurons (simple processing unit) arranged in number of layers.  The neurons in each layer feed their output forward to the next layer until we get the final output from the neural network.  There can be any number of hidden layers within a Feedforward network.  The number of neurons can be completely arbitrary. 16
  • 17. Neuron – a simple processing unit  A neuron can have any number of inputs  The inputs may be represented as x1, x2, x3… xn.  And the corresponding weights for the inputs as w1, w2, w3… wn.  Output a = x1w1+x2w2+x3w3... +xnwn 17
  • 18. • No connections within a layer • No direct connections between input and output layers • Fully connected between layers • Number of output units need not equal number of input units • Number of hidden units per layer can be more or less than input or output units y f w x bi ij j i j m = + = ( ) 1 Often include bias as an extra weight 18 Properties of architecture
  • 19. Perceptron Learning Theorem  Recap: A perceptron (threshold unit) can learn anything that it can represent (i.e. anything separable with a hyperplane) 19
  • 20. The Exclusive OR problem  A Perceptron cannot represent Exclusive OR since it is not linearly separable. 20
  • 21. 21
  • 22. Minsky & Papert (1969) offered solution to XOR problem by combining perceptron unit responses using a second layer of Units. Piecewise linear classification using an MLP with threshold (perceptron) units 1 2 +1 +1 3 22
  • 23. What do each of the layers do? 1st layer draws linear boundaries 2nd layer combines the boundaries 3rd layer can generate arbitrarily complex boundaries 23
  • 24. 2. Backward pass phase: computes ‘error signal’, propagates the error backwards through network starting at output units (where the error is the difference between actual and desired output values) 1. Forward pass phase: computes ‘functional signal’, feed forward propagation of input pattern signals through network Two phases of Backpropagation learning algorithm ‘BP’ 24 Frank Rosenblatt in 1961
  • 25. Backpropagation xk xi wki wjk dj dk yj Backward step: propagate errors from output to hidden layer Forward step: Propagate activation from input to output layer
  • 26. Backpropagation Algorithm  Initialize each wi to some small random value  Until the termination condition is met, Do  For each training example <(x1,…xn),t> Do  Input the instance (x1,…,xn) to the network and compute the network outputs yk  For each output unit k  dk=yk(1-yk)(tk-yk)  For each hidden unit h  dh=yh(1-yh) k wh,k dk  For each network weight wi,j Do  wi,j=wi,j+wi,j where wi,j=  dj xi,j
  • 27. Backpropagation  Gradient descent over entire network weight vector  Easily generalized to arbitrary directed graphs  Finds a local, not necessarily global error minimum -in practice often works well (can be invoked multiple times with different initial weights)  Often include weight momentum term wi,j(n)=  dj xi,j +  wi,j (n-1)  Minimizes error for training examples  Will it generalize well to unseen instances (out of sample error; over-fitting)?  Training can be slow typical 1000-10000 iterations
  • 28. Convergence of Backprop  Gradient descent converges to some local minimum perhaps not global minimum  Add momentum term: wki(n)  wki(n) =  dk(n) xi (n) +  wki(n-1) with   [0,1]  Stochastic gradient descent  Train multiple nets with different initial weights
  • 29. Ways to use weight derivatives  How often to update  after each training case?  after a full sweep through the training data?  How much to update  Use a fixed learning rate?  Adapt the learning rate?  Add momentum?
  • 30. Training  This was a single iteration of back-prop  Training requires many iterations with many training examples or epochs (one epoch is entire presentation of complete training set)  It can be slow !  Parallel computation implementation is also possible 30
  • 31. Training and testing data  How many examples ?  The more the merrier !  Disjoint training and testing data sets  learn from training data but evaluate performance (generalization ability) on unseen test data  Aim: minimize error on test data 31 Hoeffding’s Inequality In-sample error Out-of- sample error |H|, cardinility of hypothesis set Number of training samples