SlideShare a Scribd company logo
Artificial Intelligence
Chapter 5- Artificial Neural Networks
Content
2
 Biological Analogy
 History
 ANN Structures
 Learning in ANN
 Designing an ANN based System
 Back Propagation
 Chapter Summary
Artificial Neural Network (ANN)
3
 Biological Analogy
 ANN is a computational model that simulate some
properties of the human brain
Biological Neural Network Artificial Neural Networks
Artificial Neural Network (ANN)…
4
 Biological Analogy…
 The brain is composed of a mass of interconnected
neurons. Each neuron is connected to many other neurons
 Neurons transmit signals to each other
 Whether a signal is transmitted is an all-or-nothing event
(the electrical potential in the cell body of the neuron is
thresholded)
 Whether a signal is sent, depends on the strength of the
bond (synapse) between two neurons
Artificial Neural Network (ANN)…
5
 Brain Vs Digital Computers
 Computers require hundreds of cycles to simulate a
firing of a neuron
 The brain can fire all the neurons in a single step
o Parallelism
Parallelism
 Computers require billions of cycles to perform some
tasks but the brain takes less than a second
e.g. Face Recognition
Artificial Neural Network (ANN)…
6
 Brain Vs Digital Computers…
A crude comparison of the raw computational resources available to
computers (circa 1994) and brains
Computer Human
Computational
units
1 CPU, 105
gates 1011
neurons
Storage units 109
bits RAM, 1010
bits disk
1011
neurons, 1014
synapses
Cycle time 10-8
sec 10-3
sec
Bandwidth 109
bits/sec 1014
bits/sec
Neuron
updates/sec
105
1014
Artificial Neural Network (ANN)…
7
 History
 1943:
1943: McCulloch & Pitts show that neurons can be combined
to construct a Turing machine (using ANDs, Ors, & NOTs)
 1958:
1958: Rosenblatt shows that perceptrons will converge if
what they are trying to learn can be represented
 1969:
1969: Minsky & Papert showed the limitations of
perceptrons, killing research for a decade
 1985:
1985: backpropagation
backpropagation algorithm revitalizes the field
Artificial Neural Network (ANN)…
8
 ANN
 Is a system composed of many simple processing
elements operating in parallel which can acquire,
store, and utilize experiential knowledge
 Receives information from other neurons or from
external sources, transform the information, and
pass it on to other neurons or as external outputs
 Useful for pattern recognition, learning, and the
interpretation of incomplete inputs
Artificial Neural Network (ANN)…
9
 ANN Structures
 Feed-forward neural nets
Feed-forward neural nets:
:
 Links can only go in one direction
 Arranged in layers
layers
 Each unit is linked only in the unit in next layer
 No units are linked between the same layer, back to the
previous layer or skipping a layer
 Computations can proceed uniformly from input to output
units
 No internal state exists
Recurrent
Recurrent
Feed-forward
Feed-forward
Artificial Neural Network (ANN)…
10
 ANN Structures
 Feed-forward neural nets
Feed-forward neural nets…
…
Recurrent
Recurrent
Feed-forward
Feed-forward
Multi Layer Networks
Multi Layer Networks Perceptrons
Perceptrons
Artificial Neural Network (ANN)…
11
 ANN Structures
 Feed-forward neural nets
Feed-forward neural nets…
…
Recurrent
Recurrent
Feed-forward
Feed-forward
- Have one or more
layers of hidden units.
- With two possibly
very large hidden
layers, it is possible
possible
to implement any
to implement any
function
function.
- Networks without hidden
layer are called perceptrons.
- Perceptrons are very
limited in what they can
represent, but this makes
their learning problem much
simpler.
Multi Layer Networks
Multi Layer Networks Perceptrons
Perceptrons
Artificial Neural Network (ANN)…
12
 ANN Structures
 Recurrent neural nets
Recurrent neural nets:
:
 Links can go anywhere and form arbitrary topologies
 Allows activation to be feedback to the previous unit
 Internal state is stored in its activation level
 Can become unstable
 Can oscillate
 May take long time to compute a stable output
 Learning process is much more difficult
 Can implement more complex designs
 Can model certain systems with internal states
Recurrent
Recurrent
Feed-forward
Feed-forward
Artificial Neural Network (ANN)…
13
 Learning in ANN
 Supervised Learning
 Classification problem
 Unsupervised Learning
 Clustering
 Reinforcement Learning
 Rewarding approach
Artificial Neural Network (ANN)…
14
 Designing an ANN based system
 Get the data set
 Determine the network topology
 Training
 Optimization
 Testing
Artificial Neural Network (ANN)…
15
 Designing an ANN based system…
 Get the data set
 Example
 Determine the characteristics of the customers of a given business
institution (Bank) in terms of credit risks (low credit risk , or high credit
risk)
o The data for this problem is all about the attributes of the customers
Age
Income
Debt
Payment record
Artificial Neural Network (ANN)…
16
 Designing an ANN based system…
 Determine the network topology (Default NN topology)
 Number of nodes in the Output layer
 Equal to input parameters (independent variables)
 Number of nodes and layers in the Hidden layer
 There is no a clear cutting rule (Optimization Governs)
o Minimum error rate, and
o Less number of hidden nodes Helpful to simplify the hardware
representation of the network
 Number of nodes in the output layer
 Equal to output parameters (dependent variables)
o e.g. Number of class labels
 Possible to optimize the number of output nodes as per the nature of the
output
Artificial Neural Network (ANN)…
17
 Designing an ANN based system…
 Training
 Dataset
 Determine the training, validation, and testing dataset
 Activation function
 Back propagation algorithm
 Stopping rule
 Error rate
 Number of epoch
 Momentum coefficient
Artificial Neural Network (ANN)…
18
 Designing an ANN based system…
 Optimization: to produce an optimal NN topology
 Minimum error rate
 Less number of nodes and layers at hidden layer
 Testing
 Without input noise
 With input noise
 Using the Network
 If the error rate is within the acceptable error range, the network is ready for
use
Artificial Neural Network (ANN)…
19
 Designing an ANN based system…
 General Rules
 Initial network has a randomly assigned weights
 Learning is done by making small adjustments in the weights to
reduce the difference between the observed and predicted values
 Main difference from the logical algorithms is the need to repeat
the update phase several times in order to achieve convergence
 Updating process is divided into epochs
 Each epoch updates all the weights of the process.
Artificial Neural Network (ANN)…
20
 Designing an ANN based system…
 Example 1
 Determine the characteristics of the customers of a given business
institution (Bank) in terms of credit risks (low credit risk , or high credit
risk)
 The data for this problem is all about the attributes of the customers
o Age
o Income
o Debt
o Payment record
Artificial Neural Network (ANN)…
21
 Designing an ANN based system…
 Example 1…
Artificial Neural Network (ANN)…
22
 Backpropagation
 In 1969 a method for learning in multi-layer network,
Backpropagation
Backpropagation, was invented by Bryson and Ho
 The Backpropagation algorithm is a sensible approach
for dividing the contribution of each weight
 Works basically the same as perceptrons
Artificial Neural Network (ANN)…
23
 Backpropagation…
 Backpropagation Learning Principles: Hidden Layers
and Gradients
 There are two differences for the updating rule :
differences for the updating rule :
 The activation of the hidden unit is used instead of
instead of
the input value
the input value
 The rule contains a term for the gradient of the
activation function
Artificial Neural Network (ANN)…
24
 Backpropagation…
 Network training
 1. Initialize network with random weights (small random numbers (e.g. ranging from
-1.0 to 1.0, or -0.5 to 0.5))
 2. For all training cases (called examples):
 a. Present training inputs to network and calculate output
o Given a unit j in the hidden or output layer, the net input, Ij , to unit j is
Ij =  WijOi + өj
o Where
Wij is the weight of the connection from unit i in the previous layer to unit j;
Oi is the output of unit i from the previous layer ; and
Өj is the bias of the unit
 b.
b. activation function: Given the net input Ij to unit j, then Oj , the out put of unit j is
computed as
Oj = 1 / (1 + e-Ij
) ---------- Sigmoid function
Artificial Neural Network (ANN)…
25
 Backpropagation…
 Network training…
 3. Back propagate the error
 For a unit j in the output layer , the error Errj is computed by
Errj = Oj(1- Oj ) (Tj- Oj )
o Where
Oj is the actual output of unit j
Tj is true output Oj
Oj(1- Oj ) the deviation of logistic function
 To compute the error of a hidden layer unit j , the weighted sum of the errors of the
units connected to unit j in the next layer are considered. The error of the hidden layer
unit j is
Errj = Oj(1- Oj )  ErrkWjk
o Where
Wjk is weight of the connection from unit j to a unit k in the next higher layer
Errk is the error of unit k
Artificial Neural Network (ANN)…
26
 Backpropagation…
 Network training…
 4. Update weights and biases to reflect the propagated error
 Update weights
ΔWij = (l)ErrjOi
Wij = Wij + ΔWij
o Where
ΔWij the change in weight Wij
l learning rate (a constant typically having the value b/n 0.0 and 1.0)
(the rule of thumb is to set the learning rate to 1/t, where t is the number
of iterations through the training set so far.
 Update biases
Δ өj = (l)Errj
өj = өj + Δ өj
Artificial Neural Network (ANN)…
27
 Backpropagation…
 Network training…
 5. Terminating conditions
 All ΔWij =in the previous epoch were so small as to be below
some specified threshold, or
 The percentage of samples misclassified in the previous epoch
is below some threshold, or
 A prespecified number of epochs has expired
Artificial Neural Network (ANN)…
28
 Backpropagation…
 Example 1
4
2
3
5
6
x1
x2
x3
w14
w15
w24
w25
w34
w35
w46
w56
The above figure shows a multilayer feed-forward neural network. Let the
learning rate be 0.9. The initial weight and bias values of the network
are given in the table along with the first training sample, X = (1, 0, 1),
whose class label is 1. Show the calculation for the backpropagation.
x1 x2 x3 w14 w15 w24 w25 w34 w35 w46 w56 ө4 ө5 ө6
1 0 1 0.2 -0.3 0.4 0.1 -0.5 0.2 -0.3 -0.2 -0.2 0.2 0.1
Artificial Neural Network (ANN)…
29
 Summary
 NNs are used for classification and function approximation
or mapping problems which are:
 Tolerant of some imprecision
 Have lots of training data available
 Hard and fast rules cannot easily be applied
Artificial Neural Network (ANN)…
30
 Summary…
 ANNs applications can be grouped in to the following categories:
 Clustering:
 A clustering algorithm explores the similarity between patterns and
places similar patterns in a cluster. Best known applications include data
compression and data mining.
 Classification/Pattern recognition:
 The task of pattern recognition is to assign an input pattern (like
handwritten symbol) to one of many classes. This category includes
algorithmic implementations such as associative memory.
 Function approximation:
 The tasks of function approximation is to find an estimate of the unknown
function f() subject to noise. Various engineering and scientific disciplines
require function approximation
Artificial Neural Network (ANN)…
31
 Summary…
 ANNs applications can be grouped in to the following
categories:…
 Prediction/Dynamical Systems:
 The task is to forecast some future values of a time-sequenced
data. Prediction has a significant impact on decision support
systems. Prediction differs from Function approximation by
considering time factor. Here the system is dynamic and may
produce different results for the same input data based on system
state (time).
Assignment-2
32
 Explain the following terms (with in the perspective of ANNs)
 Epoch
 Learning rate
 Momentum coefficient
 Explain sampling techniques used to determine the training, validation, and
testing datasets in the training of ANNs.
 Compare ANN with
 Support Vector Machine (SVM)
 Decision tree
 Bayesian Belief Network
Due date 04.01.2018
Project
33
 Consider a classification problem in a supervised
environment and develop an ANN based expert
systems in the domains given below
 Group-1
 Agriculture
 Group-2
 Energy
 Group-3
 Health
 Report submission and Presentation due date
 26.01.18

More Related Content

PPTX
PPTX
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
PPT
SET-02_SOCS_ESE-DEC23__B.Tech%20(CSE-H+NH)-AIML_5_CSAI300
PPTX
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
PPTX
employed to cover the tampering traces of a tampered image. Image tampering
PPT
ANNs have been widely used in various domains for: Pattern recognition Funct...
PPT
ai...........................................
Jyduydufyuyf8yfiyfiyfifiyfiyviyviyfiugiuy8f7dd64d4yrsxyfhgdhfjhvjhv
SET-02_SOCS_ESE-DEC23__B.Tech%20(CSE-H+NH)-AIML_5_CSAI300
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
employed to cover the tampering traces of a tampered image. Image tampering
ANNs have been widely used in various domains for: Pattern recognition Funct...
ai...........................................

Similar to AI-CH5 (ANN) - Artificial Neural Network (20)

PPT
ai7 (1) Artificial Neural Network Intro .ppt
PPT
ann ppt , multilayer perceptron. presentation on
PPT
Game theory.pdf textbooks content Artificical
PPT
Machine Learning Neural Networks Artificial Intelligence
PPT
Machine Learning Neural Networks Artificial
PPT
INTRODUCTION TO ARTIFICIAL INTELLIGENCE.
PPT
ann-ics320Part4.ppt
PPT
ann-ics320Part4.ppt
PPTX
Artificial Neural Networks ppt.pptx for final sem cse
PPTX
ML_Unit_2_Part_A
PPT
ai7.ppt
PPT
ai7.ppt
PPT
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
PPTX
Chapter-5-Part I-Basics-Neural-Networks.pptx
PPTX
02 Fundamental Concepts of ANN
PPTX
Artificial neural networks
PPTX
10 Backpropagation Algorithm for Neural Networks (1).pptx
PPT
Neural-Networks.ppt
PPTX
Artifical Neural Network
PPTX
Artificial Neural Networks presentations
ai7 (1) Artificial Neural Network Intro .ppt
ann ppt , multilayer perceptron. presentation on
Game theory.pdf textbooks content Artificical
Machine Learning Neural Networks Artificial Intelligence
Machine Learning Neural Networks Artificial
INTRODUCTION TO ARTIFICIAL INTELLIGENCE.
ann-ics320Part4.ppt
ann-ics320Part4.ppt
Artificial Neural Networks ppt.pptx for final sem cse
ML_Unit_2_Part_A
ai7.ppt
ai7.ppt
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
Chapter-5-Part I-Basics-Neural-Networks.pptx
02 Fundamental Concepts of ANN
Artificial neural networks
10 Backpropagation Algorithm for Neural Networks (1).pptx
Neural-Networks.ppt
Artifical Neural Network
Artificial Neural Networks presentations
Ad

Recently uploaded (20)

PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
PDF
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PPT
Quality review (1)_presentation of this 21
PPTX
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
PPTX
Introduction to Knowledge Engineering Part 1
PDF
Clinical guidelines as a resource for EBP(1).pdf
PPT
Miokarditis (Inflamasi pada Otot Jantung)
PPTX
01_intro xxxxxxxxxxfffffffffffaaaaaaaaaaafg
PPTX
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
PDF
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
PPTX
IBA_Chapter_11_Slides_Final_Accessible.pptx
PDF
Fluorescence-microscope_Botany_detailed content
PDF
Lecture1 pattern recognition............
PDF
.pdf is not working space design for the following data for the following dat...
PPT
Reliability_Chapter_ presentation 1221.5784
PPTX
oil_refinery_comprehensive_20250804084928 (1).pptx
Data_Analytics_and_PowerBI_Presentation.pptx
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
Recruitment and Placement PPT.pdfbjfibjdfbjfobj
Business Ppt On Nestle.pptx huunnnhhgfvu
Acceptance and paychological effects of mandatory extra coach I classes.pptx
Galatica Smart Energy Infrastructure Startup Pitch Deck
Quality review (1)_presentation of this 21
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
Introduction to Knowledge Engineering Part 1
Clinical guidelines as a resource for EBP(1).pdf
Miokarditis (Inflamasi pada Otot Jantung)
01_intro xxxxxxxxxxfffffffffffaaaaaaaaaaafg
MODULE 8 - DISASTER risk PREPAREDNESS.pptx
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
IBA_Chapter_11_Slides_Final_Accessible.pptx
Fluorescence-microscope_Botany_detailed content
Lecture1 pattern recognition............
.pdf is not working space design for the following data for the following dat...
Reliability_Chapter_ presentation 1221.5784
oil_refinery_comprehensive_20250804084928 (1).pptx
Ad

AI-CH5 (ANN) - Artificial Neural Network

  • 1. Artificial Intelligence Chapter 5- Artificial Neural Networks
  • 2. Content 2  Biological Analogy  History  ANN Structures  Learning in ANN  Designing an ANN based System  Back Propagation  Chapter Summary
  • 3. Artificial Neural Network (ANN) 3  Biological Analogy  ANN is a computational model that simulate some properties of the human brain Biological Neural Network Artificial Neural Networks
  • 4. Artificial Neural Network (ANN)… 4  Biological Analogy…  The brain is composed of a mass of interconnected neurons. Each neuron is connected to many other neurons  Neurons transmit signals to each other  Whether a signal is transmitted is an all-or-nothing event (the electrical potential in the cell body of the neuron is thresholded)  Whether a signal is sent, depends on the strength of the bond (synapse) between two neurons
  • 5. Artificial Neural Network (ANN)… 5  Brain Vs Digital Computers  Computers require hundreds of cycles to simulate a firing of a neuron  The brain can fire all the neurons in a single step o Parallelism Parallelism  Computers require billions of cycles to perform some tasks but the brain takes less than a second e.g. Face Recognition
  • 6. Artificial Neural Network (ANN)… 6  Brain Vs Digital Computers… A crude comparison of the raw computational resources available to computers (circa 1994) and brains Computer Human Computational units 1 CPU, 105 gates 1011 neurons Storage units 109 bits RAM, 1010 bits disk 1011 neurons, 1014 synapses Cycle time 10-8 sec 10-3 sec Bandwidth 109 bits/sec 1014 bits/sec Neuron updates/sec 105 1014
  • 7. Artificial Neural Network (ANN)… 7  History  1943: 1943: McCulloch & Pitts show that neurons can be combined to construct a Turing machine (using ANDs, Ors, & NOTs)  1958: 1958: Rosenblatt shows that perceptrons will converge if what they are trying to learn can be represented  1969: 1969: Minsky & Papert showed the limitations of perceptrons, killing research for a decade  1985: 1985: backpropagation backpropagation algorithm revitalizes the field
  • 8. Artificial Neural Network (ANN)… 8  ANN  Is a system composed of many simple processing elements operating in parallel which can acquire, store, and utilize experiential knowledge  Receives information from other neurons or from external sources, transform the information, and pass it on to other neurons or as external outputs  Useful for pattern recognition, learning, and the interpretation of incomplete inputs
  • 9. Artificial Neural Network (ANN)… 9  ANN Structures  Feed-forward neural nets Feed-forward neural nets: :  Links can only go in one direction  Arranged in layers layers  Each unit is linked only in the unit in next layer  No units are linked between the same layer, back to the previous layer or skipping a layer  Computations can proceed uniformly from input to output units  No internal state exists Recurrent Recurrent Feed-forward Feed-forward
  • 10. Artificial Neural Network (ANN)… 10  ANN Structures  Feed-forward neural nets Feed-forward neural nets… … Recurrent Recurrent Feed-forward Feed-forward Multi Layer Networks Multi Layer Networks Perceptrons Perceptrons
  • 11. Artificial Neural Network (ANN)… 11  ANN Structures  Feed-forward neural nets Feed-forward neural nets… … Recurrent Recurrent Feed-forward Feed-forward - Have one or more layers of hidden units. - With two possibly very large hidden layers, it is possible possible to implement any to implement any function function. - Networks without hidden layer are called perceptrons. - Perceptrons are very limited in what they can represent, but this makes their learning problem much simpler. Multi Layer Networks Multi Layer Networks Perceptrons Perceptrons
  • 12. Artificial Neural Network (ANN)… 12  ANN Structures  Recurrent neural nets Recurrent neural nets: :  Links can go anywhere and form arbitrary topologies  Allows activation to be feedback to the previous unit  Internal state is stored in its activation level  Can become unstable  Can oscillate  May take long time to compute a stable output  Learning process is much more difficult  Can implement more complex designs  Can model certain systems with internal states Recurrent Recurrent Feed-forward Feed-forward
  • 13. Artificial Neural Network (ANN)… 13  Learning in ANN  Supervised Learning  Classification problem  Unsupervised Learning  Clustering  Reinforcement Learning  Rewarding approach
  • 14. Artificial Neural Network (ANN)… 14  Designing an ANN based system  Get the data set  Determine the network topology  Training  Optimization  Testing
  • 15. Artificial Neural Network (ANN)… 15  Designing an ANN based system…  Get the data set  Example  Determine the characteristics of the customers of a given business institution (Bank) in terms of credit risks (low credit risk , or high credit risk) o The data for this problem is all about the attributes of the customers Age Income Debt Payment record
  • 16. Artificial Neural Network (ANN)… 16  Designing an ANN based system…  Determine the network topology (Default NN topology)  Number of nodes in the Output layer  Equal to input parameters (independent variables)  Number of nodes and layers in the Hidden layer  There is no a clear cutting rule (Optimization Governs) o Minimum error rate, and o Less number of hidden nodes Helpful to simplify the hardware representation of the network  Number of nodes in the output layer  Equal to output parameters (dependent variables) o e.g. Number of class labels  Possible to optimize the number of output nodes as per the nature of the output
  • 17. Artificial Neural Network (ANN)… 17  Designing an ANN based system…  Training  Dataset  Determine the training, validation, and testing dataset  Activation function  Back propagation algorithm  Stopping rule  Error rate  Number of epoch  Momentum coefficient
  • 18. Artificial Neural Network (ANN)… 18  Designing an ANN based system…  Optimization: to produce an optimal NN topology  Minimum error rate  Less number of nodes and layers at hidden layer  Testing  Without input noise  With input noise  Using the Network  If the error rate is within the acceptable error range, the network is ready for use
  • 19. Artificial Neural Network (ANN)… 19  Designing an ANN based system…  General Rules  Initial network has a randomly assigned weights  Learning is done by making small adjustments in the weights to reduce the difference between the observed and predicted values  Main difference from the logical algorithms is the need to repeat the update phase several times in order to achieve convergence  Updating process is divided into epochs  Each epoch updates all the weights of the process.
  • 20. Artificial Neural Network (ANN)… 20  Designing an ANN based system…  Example 1  Determine the characteristics of the customers of a given business institution (Bank) in terms of credit risks (low credit risk , or high credit risk)  The data for this problem is all about the attributes of the customers o Age o Income o Debt o Payment record
  • 21. Artificial Neural Network (ANN)… 21  Designing an ANN based system…  Example 1…
  • 22. Artificial Neural Network (ANN)… 22  Backpropagation  In 1969 a method for learning in multi-layer network, Backpropagation Backpropagation, was invented by Bryson and Ho  The Backpropagation algorithm is a sensible approach for dividing the contribution of each weight  Works basically the same as perceptrons
  • 23. Artificial Neural Network (ANN)… 23  Backpropagation…  Backpropagation Learning Principles: Hidden Layers and Gradients  There are two differences for the updating rule : differences for the updating rule :  The activation of the hidden unit is used instead of instead of the input value the input value  The rule contains a term for the gradient of the activation function
  • 24. Artificial Neural Network (ANN)… 24  Backpropagation…  Network training  1. Initialize network with random weights (small random numbers (e.g. ranging from -1.0 to 1.0, or -0.5 to 0.5))  2. For all training cases (called examples):  a. Present training inputs to network and calculate output o Given a unit j in the hidden or output layer, the net input, Ij , to unit j is Ij =  WijOi + өj o Where Wij is the weight of the connection from unit i in the previous layer to unit j; Oi is the output of unit i from the previous layer ; and Өj is the bias of the unit  b. b. activation function: Given the net input Ij to unit j, then Oj , the out put of unit j is computed as Oj = 1 / (1 + e-Ij ) ---------- Sigmoid function
  • 25. Artificial Neural Network (ANN)… 25  Backpropagation…  Network training…  3. Back propagate the error  For a unit j in the output layer , the error Errj is computed by Errj = Oj(1- Oj ) (Tj- Oj ) o Where Oj is the actual output of unit j Tj is true output Oj Oj(1- Oj ) the deviation of logistic function  To compute the error of a hidden layer unit j , the weighted sum of the errors of the units connected to unit j in the next layer are considered. The error of the hidden layer unit j is Errj = Oj(1- Oj )  ErrkWjk o Where Wjk is weight of the connection from unit j to a unit k in the next higher layer Errk is the error of unit k
  • 26. Artificial Neural Network (ANN)… 26  Backpropagation…  Network training…  4. Update weights and biases to reflect the propagated error  Update weights ΔWij = (l)ErrjOi Wij = Wij + ΔWij o Where ΔWij the change in weight Wij l learning rate (a constant typically having the value b/n 0.0 and 1.0) (the rule of thumb is to set the learning rate to 1/t, where t is the number of iterations through the training set so far.  Update biases Δ өj = (l)Errj өj = өj + Δ өj
  • 27. Artificial Neural Network (ANN)… 27  Backpropagation…  Network training…  5. Terminating conditions  All ΔWij =in the previous epoch were so small as to be below some specified threshold, or  The percentage of samples misclassified in the previous epoch is below some threshold, or  A prespecified number of epochs has expired
  • 28. Artificial Neural Network (ANN)… 28  Backpropagation…  Example 1 4 2 3 5 6 x1 x2 x3 w14 w15 w24 w25 w34 w35 w46 w56 The above figure shows a multilayer feed-forward neural network. Let the learning rate be 0.9. The initial weight and bias values of the network are given in the table along with the first training sample, X = (1, 0, 1), whose class label is 1. Show the calculation for the backpropagation. x1 x2 x3 w14 w15 w24 w25 w34 w35 w46 w56 ө4 ө5 ө6 1 0 1 0.2 -0.3 0.4 0.1 -0.5 0.2 -0.3 -0.2 -0.2 0.2 0.1
  • 29. Artificial Neural Network (ANN)… 29  Summary  NNs are used for classification and function approximation or mapping problems which are:  Tolerant of some imprecision  Have lots of training data available  Hard and fast rules cannot easily be applied
  • 30. Artificial Neural Network (ANN)… 30  Summary…  ANNs applications can be grouped in to the following categories:  Clustering:  A clustering algorithm explores the similarity between patterns and places similar patterns in a cluster. Best known applications include data compression and data mining.  Classification/Pattern recognition:  The task of pattern recognition is to assign an input pattern (like handwritten symbol) to one of many classes. This category includes algorithmic implementations such as associative memory.  Function approximation:  The tasks of function approximation is to find an estimate of the unknown function f() subject to noise. Various engineering and scientific disciplines require function approximation
  • 31. Artificial Neural Network (ANN)… 31  Summary…  ANNs applications can be grouped in to the following categories:…  Prediction/Dynamical Systems:  The task is to forecast some future values of a time-sequenced data. Prediction has a significant impact on decision support systems. Prediction differs from Function approximation by considering time factor. Here the system is dynamic and may produce different results for the same input data based on system state (time).
  • 32. Assignment-2 32  Explain the following terms (with in the perspective of ANNs)  Epoch  Learning rate  Momentum coefficient  Explain sampling techniques used to determine the training, validation, and testing datasets in the training of ANNs.  Compare ANN with  Support Vector Machine (SVM)  Decision tree  Bayesian Belief Network Due date 04.01.2018
  • 33. Project 33  Consider a classification problem in a supervised environment and develop an ANN based expert systems in the domains given below  Group-1  Agriculture  Group-2  Energy  Group-3  Health  Report submission and Presentation due date  26.01.18