SlideShare a Scribd company logo
Artificial Neural Network (ANN)
A. Introduction to neural networks
B. ANN architectures
• Feedforward networks
• Feedback networks
• Lateral networks
C. Learning methods
• Supervised learning
• Unsupervised learning
• Reinforced learning
D. Learning rule on supervised learning
• Gradient descent,
• Widrow-hoff (LMS)
• Generalized delta
• Error-correction
E. Feedforward neural network with Gradient descent optimization
Introduction to neural networks
Definition: the ability to learn, memorize and still
generalize, prompted research in algorithmic
modeling of biological neural systems
Do you think that computer smarter than human
brain?
“While successes have been achieved in modeling biological neural systems, there are still no
“While successes have been achieved in modeling biological neural systems, there are still no
solutions to the complex problem of modeling intuition, consciousness and emotion
solutions to the complex problem of modeling intuition, consciousness and emotion -
- which
which
form
form integral parts of human intelligence”…(
integral parts of human intelligence”…(Alan Turing, 1950)
---Human brain has the ability to perform tasks such as pattern recognition,
perception and motor control much faster than any computer---
Facts of Human Brain
(complex, nonlinear and parallel computer)
• The brain contains about 1010 (100
billion) basic units called neurons
• Each neuron connected to about 104
other neurons
• Weight: birth 0.3 kg, adult ~1.5 kg
• Power consumption 20-40W (~20%
of body consumption)
• Signal propagation speed inside the
axon ~90m/s in ~170,000 Km of axon
length for adult male
• Firing frequency of a neuron ~250 –
2000Hz
• Operating temperature: 37±2oC
• Sleep requirement: average 7.5 hours
(adult)
Intel Pentium 4 1.5GHz
Number of transistors 4.2x107
Power consumption up to 55 Watts
Weight
0.1 kg cartridge w/o
fans, 0.3 kg with
fan/heatsink
Maximum firing
frequency
1.5 GHz
Normal operating
temperature
15-85°C
Sleep requirement
0 (if not overheated/
overclocked)
Processing of complex
stimuli
if can be done, takes a
long time
Biological neuron
• Soma: Nucleus of neuron (the cell body) -
process the input
• Dendrites: long irregularly shaped filaments
attached to the soma – input channels
• Axon: another type link attached to the
soma – output channels
• Output of the axon: voltage pulse (spike)
voltage pulse (spike)
that lasts for a ms
• Firing of neuron – membrane potential
• Axon terminates in a specialized contact
called the synaptic junction – the
electrochemical contact between neurons
• The size of synapses are believed to be
linked with learning
• Larger area: excitatory—smaller area:
inhibitory
Artificial neuron model
(McCulloh-Pitts model, 1949)
Qj : external threshold, offset or bias
wji : synaptic weights
xi : input
yj : output
…..Another model-Product unit
Firing and the strength of the exiting signal
are controlled by activation function (AF)
Allow higher-order combinations of inputs, having the advantage of
increased information capacity
Types of AF:
•
•Linear
Linear
•
•Step
Step
•
•Ramp
Ramp
•
•Sigmoid
Sigmoid
•
•Hyperbolic tangent
Hyperbolic tangent
•
•Gaussian
Gaussian
Different NN types
• Single-layer NNs, such as the Hopfield network
• Multilayer feedforward NNs, for example standard
backpropagation, functional link and product unit networks
• Temporal NNs, such as the Elman and Jordan simple recurrent
networks as well as time-delay neural networks
• Self-organizing NNs, such as the Kohonen self-organizing
feature maps and the learning vector quantizer
• Combined feedforward and self-organizing NNs, such as the
radial basis function networks
The ANN applications
•
• Classification,
Classification, the aim is to predict the class of an input vector
•
• Pattern matching
Pattern matching, the aim is to produce a pattern best associated with a
given input vector
•
• Pattern completion
Pattern completion, the aim is to complete the missing parts of a given
input vector
•
• Optimization
Optimization, the aim is to find the optimal values of parameters in an
optimization problem
•
• Control
Control, an appropriate action is suggested based on given an input
vectors
•
• Function approximation/times series modeling
Function approximation/times series modeling, the aim is to learn the
functional relationships between input and desired output vectors;
•
• Data mining
Data mining, with the aim of discovering hidden patterns from data
(knowledge discovery)
ANN architectures
• Neural Networks are known to be universal function
approximators
• Various architectures are available to approximate any
nonlinear function
• Different architectures allow for generation of functions of
different complexity and power
Feedforward networks
Feedback networks
Lateral networks
Feedforward Networks
Network size: n x m x r = 2x5x1
Wmn: input weight matrix
Vrm: output weight matrix
•No feedback within the network
•The coupling takes place from one layer to the next
•The information flows, in general, in the forward
direction
Input layer: Number of neurons in this
layer corresponds to the number of
inputs to the neuronal network. This
layer consists of passive nodes, i.e.,
which do not take part in the actual
signal modification, but only transmits
the signal to the following layer.
• Hidden layer: This layer has arbitrary
number of layers with arbitrary number
of neurons. The nodes in this layer take
part in the signal modification, hence,
they are active.
• Output layer: The number of neurons
in the output layer corresponds to the
number of the output values of the
neural network. The nodes in this layer
are active ones.
FFNN can have more than one hidden layer.
However, it has been proved that FFNNs with
one hidden layer has enough to approximate
any continuous function [Hornik 1989].
Feedback networks
Elman Recurrent Network
The output of a neuron is either directly or indirectly
fed back to its input via other linked neurons used
in complex pattern recognition tasks, e.g., speech
recognition etc.
Feedback networks
Jordan Recurrent Network
Lateral Networks
•There exist couplings of neurons within one layer
•There is no essentially explicit feedback path amongst the different layers
•This can be thought of as a compromise between the forward and feedback
network
Input layer Hidden layer Output layer
Learning methods
• Artificial neural networks work through the optimized weight values.
• The method by which the optimized weight values are attained is called
learning
learning
• In the learning process  try to teach the network how to produce the
output when the corresponding input is presented
• When learning is complete: the trained neural network, with the updated
optimal weights, should be able to produce the output within desired
accuracy corresponding to an input pattern.
Learning methods
• Supervised learning
• Unsupervised learning
• Reinforced learning
Classification of Learning Algorithms
Supervised learning
Supervised learning means guided learning by
“teacher”; requires a training set which consists
of input vectors and a target vector associated
with each input vector
“Learning experience in our childhood”
As a child, we learn about various things
(input) when we see them and
simultaneously are told (supervised)
about their names and the respective
functionalities (desired response).
Supervised learning system:
•feedforward
•functional link
•product unit
•Recurrent
•Time delay
Unsupervised learning
• The objective of unsupervised learning is to discover patterns or features
in the input data with no help from a teacher, basically performing a
clustering of input space.
• The system learns about the pattern from the data itself without a priori
knowledge. This is similar to our learning experience in adulthood
“For example, often in our working environment we are thrown into a
project or situation which we know very little about. However, we try to
familiarize with the situation as quickly as possible using our previous
experiences, education, willingness and similar other factors”
• Hebb’s rule: It helps the neural network or neuron assemblies to
remember specific patterns much like the memory. From that stored
knowledge, similar sort of incomplete or spatial patterns could be
recognized. This is even faster than the delta rule or the backpropagation
algorithm because there is no repetitive presentation and training of
input–output pairs.
Reinforced learning
•A ‘teacher’ though available, does not present the expected answer but only indicates if
the computed output is correct or incorrect
•The information provided helps the network in its learning process
•A reward is given for a correct answer computed and a penalty for a wrong answer
Leaning algorithm in
Supervised learning
• Gradient descent
• Widrow-hoff (LMS)
• Generalized delta
• Error-correction
Single neuron
Gradient Descent
• Gradient descent (GD)…(not the first but used most)
• GD is aimed to find the weight values that minimize Error
• GD requires the definition of an error (or objective)
function to measure the neuron's error in approximating
the target
Analogy: Suppose we want to come down
(descend) from a high hill (higher error) to a
low valley (lower error). We move along the
negative gradient or slopes. By doing so, we
take the steepest path to the downhill
valley steepest descent algorithm
Where tp and fp are respectively the target and actual output for patterns p
The updated weights:
where η :learning rate
wi(t+1):new weights
The calculation of the partial derivative of f
with respect to up (the net input for pattern p)
presents a problem for all discontinuous
activation functions,
such as the step and ramp functions
Widrow-Hoff learning rule
Widrow-hoff
Least-Means-Square (LMS)
Assume that f = up
The weights are updated using:
One of the first algorithms used to train multiple adaptive linear neurons
(Madaline) [Widrow 1987, Widrow and Lehr 1990]
Generalized delta
Assume: differentiable activation functions; such as sigmoid function
The weights are updated using:
Error-correction
Assume that binary-valued functions are used, e.g the step function.
The weights are updated using:
Weights are only adjusted when the neuron responds in error
Feedforward neural network with
Gradient descent optimization
Input vectors  actual value is calculated
then error is calculated
The error gradient respect to network’s weight is calculated
by propagating the error backward through network
Once the error gradient is calculated, the weight is adjusted
More details…
Functional Diagram of FFNN
Feedforward Operation
Input vector xj where j =1 to n (number of inputs)
Input weight matrix Wij where i = 1 to m (hidden neurons)
Step 1: Activation vector ai :
Decision vector di :
Step 2: Output vector yi is given by
(r is no. of outputs):

Backpropagation Operation
Step 1: The output error vector:
Step 2: The decision error vector:
The activation error vector:
Step 3: The weights changes:
gg and gm are learning and momentum rates, respectively
The weights updates:
One set of weight modifications is called an epoch, and
many of these may be required before the desired
accuracy of approximation is reached.
“This is the objective function for NN learning that need
to be optimized
optimized by the optimization methods”
The backpropagation training algorithm is based on
the principle of gradient descent and is given as half
the square of the Euclidean norm of the output
error vector.
Optimization methods to carry out NN learning
• Local optimization, where the algorithm ends up in a
local optimum without finding a global optimum.
Gradient descent and scaled conjugate gradient are
local optimizers.
• Global optimization, where the algorithm searches
for the global optimum by with mechanisms that
allow greater search space explorations. Global
optimizers include Leapfrog, simulated annealing,
evolutionary computing and swarm optimization.
“Local and global optimization techniques can be combined to form
hybrid training algorithms”
Weight Adjustments/Updates
• Stochastic/Delta/(online) learning, where the NN weights are adjusted
after each pattern presentation. In this case the next input pattern is
selected randomly from the training set, to prevent any bias that may
occur due to the sequences in which patterns occur in the training set.
• Batch/(offline) learning, where the NN weight changes are accumulated
and used to adjust weights only after all training patterns have been
presented
Two types of supervised learning algorithms exist, based on when/how weights
are updated:
Feedforward Neural Networks
(effects of weight variations)
Homework!
• Please make a report about the potential of
intelligent techniques is applied for the part of
your current research
• Due date: 12 May 2010
Project!
Critical review
Please find very recent paper about the application of intelligent techniques
in power system area; try to understand the paper and criticize
the contents
Due date: will be at the final lecture

More Related Content

PPTX
Reason To Switch to DNNDNNs excel in handling huge volumes of data (e.g., ima...
PPTX
Introduction to Neural networks (under graduate course) Lecture 9 of 9
PPTX
Deep Learning Sample Class (Jon Lederman)
PPTX
02 Fundamental Concepts of ANN
PPTX
neuralnetwork.pptx
PPTX
neuralnetwork.pptx
PPT
Artificial neural network
PPTX
Neural Networks and Deep Learning Basics
Reason To Switch to DNNDNNs excel in handling huge volumes of data (e.g., ima...
Introduction to Neural networks (under graduate course) Lecture 9 of 9
Deep Learning Sample Class (Jon Lederman)
02 Fundamental Concepts of ANN
neuralnetwork.pptx
neuralnetwork.pptx
Artificial neural network
Neural Networks and Deep Learning Basics

Similar to Artificial Neural Network (ANN (20)

PDF
Deep learning unit 3 artificial neural network
PPTX
Neural network
PPT
neuralnetworklearningalgorithm-231219123006-bb13a863.ppt
PDF
33.-Multi-Layer-Perceptron.pdf
PDF
Artificial Neural Networks: Introduction, Neural Network representation, Appr...
PPTX
Introduction to Deep Leaning(UNIT 1).pptx
PPT
Artificial Neural Network Learning Algorithm.ppt
PPTX
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
PPTX
Artificial neural network
PDF
Machine Learning: Introduction to Neural Networks
PDF
Introduction to Artificial Intelligence_ Lec 9
PPTX
Artificial Neural Network
PPTX
Terminology Machine Learning
PPTX
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
PPTX
employed to cover the tampering traces of a tampered image. Image tampering
PDF
Competitive Neural Network Artificial Intelligence
PPTX
NN and DL_Intro__ to Neural Network.pptx
PPTX
UNIT-II_5.Neural Networks.pptx Artificial neural networks
PPT
neuros
PPT
Neutral Network
Deep learning unit 3 artificial neural network
Neural network
neuralnetworklearningalgorithm-231219123006-bb13a863.ppt
33.-Multi-Layer-Perceptron.pdf
Artificial Neural Networks: Introduction, Neural Network representation, Appr...
Introduction to Deep Leaning(UNIT 1).pptx
Artificial Neural Network Learning Algorithm.ppt
NEURAL NETWORK IN MACHINE LEARNING FOR STUDENTS
Artificial neural network
Machine Learning: Introduction to Neural Networks
Introduction to Artificial Intelligence_ Lec 9
Artificial Neural Network
Terminology Machine Learning
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
employed to cover the tampering traces of a tampered image. Image tampering
Competitive Neural Network Artificial Intelligence
NN and DL_Intro__ to Neural Network.pptx
UNIT-II_5.Neural Networks.pptx Artificial neural networks
neuros
Neutral Network
Ad

More from Andrew Molina (20)

PDF
50 Self Evaluation Examples, Forms Questions Templat
PDF
How To Make An Intro Paragraph. Intr
PDF
Essay Websites My Favorite Ho
PDF
Writing Papers High - The Best Place To Buy Same D
PDF
New YearS Writing Paper Writing Paper, Lined Paper
PDF
How To Write A Paper On A Topic ~ Abbeye Booklet
PDF
How To Write An Essay About Myself Myself Handwritin
PDF
School Essay Writing A Paper Proposal
PDF
Tooth Fairy Letter Template Free - Lassafolder
PDF
Features Of A Play Script Display Teaching Resources
PDF
5-Paragraph Essay Graphic Organizer - In Our Pond
PDF
How To Write A Research Paper Middle School Powerpoint - School
PDF
Essay Exams Essays Test (Assessment)
PDF
What Does 500 Words Look Like 500 Word Essay, Ess
PDF
Marking Rubric Essay, Essay Writing, Good Essay
PDF
40 Case Brief Examples Templates TemplateLab
PDF
Examples Of Argument Essays Essay Examples,
PDF
Descriptive Essay Gradesaver Bu
PDF
I Need Someone To Write My E
PDF
Resume Paper Best Types, Colors Brands To Choos
50 Self Evaluation Examples, Forms Questions Templat
How To Make An Intro Paragraph. Intr
Essay Websites My Favorite Ho
Writing Papers High - The Best Place To Buy Same D
New YearS Writing Paper Writing Paper, Lined Paper
How To Write A Paper On A Topic ~ Abbeye Booklet
How To Write An Essay About Myself Myself Handwritin
School Essay Writing A Paper Proposal
Tooth Fairy Letter Template Free - Lassafolder
Features Of A Play Script Display Teaching Resources
5-Paragraph Essay Graphic Organizer - In Our Pond
How To Write A Research Paper Middle School Powerpoint - School
Essay Exams Essays Test (Assessment)
What Does 500 Words Look Like 500 Word Essay, Ess
Marking Rubric Essay, Essay Writing, Good Essay
40 Case Brief Examples Templates TemplateLab
Examples Of Argument Essays Essay Examples,
Descriptive Essay Gradesaver Bu
I Need Someone To Write My E
Resume Paper Best Types, Colors Brands To Choos
Ad

Recently uploaded (20)

PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PPTX
Pharmacology of Heart Failure /Pharmacotherapy of CHF
PDF
Classroom Observation Tools for Teachers
PDF
Complications of Minimal Access Surgery at WLH
PDF
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PDF
Chapter 2 Heredity, Prenatal Development, and Birth.pdf
PPTX
Renaissance Architecture: A Journey from Faith to Humanism
PDF
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
PDF
Insiders guide to clinical Medicine.pdf
PDF
Module 4: Burden of Disease Tutorial Slides S2 2025
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
PDF
01-Introduction-to-Information-Management.pdf
PDF
2.FourierTransform-ShortQuestionswithAnswers.pdf
PPTX
PPH.pptx obstetrics and gynecology in nursing
PPTX
Week 4 Term 3 Study Techniques revisited.pptx
PDF
TR - Agricultural Crops Production NC III.pdf
PDF
Supply Chain Operations Speaking Notes -ICLT Program
FourierSeries-QuestionsWithAnswers(Part-A).pdf
Pharmacology of Heart Failure /Pharmacotherapy of CHF
Classroom Observation Tools for Teachers
Complications of Minimal Access Surgery at WLH
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
Chapter 2 Heredity, Prenatal Development, and Birth.pdf
Renaissance Architecture: A Journey from Faith to Humanism
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
Insiders guide to clinical Medicine.pdf
Module 4: Burden of Disease Tutorial Slides S2 2025
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
STATICS OF THE RIGID BODIES Hibbelers.pdf
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
01-Introduction-to-Information-Management.pdf
2.FourierTransform-ShortQuestionswithAnswers.pdf
PPH.pptx obstetrics and gynecology in nursing
Week 4 Term 3 Study Techniques revisited.pptx
TR - Agricultural Crops Production NC III.pdf
Supply Chain Operations Speaking Notes -ICLT Program

Artificial Neural Network (ANN

  • 1. Artificial Neural Network (ANN) A. Introduction to neural networks B. ANN architectures • Feedforward networks • Feedback networks • Lateral networks C. Learning methods • Supervised learning • Unsupervised learning • Reinforced learning D. Learning rule on supervised learning • Gradient descent, • Widrow-hoff (LMS) • Generalized delta • Error-correction E. Feedforward neural network with Gradient descent optimization
  • 2. Introduction to neural networks Definition: the ability to learn, memorize and still generalize, prompted research in algorithmic modeling of biological neural systems Do you think that computer smarter than human brain? “While successes have been achieved in modeling biological neural systems, there are still no “While successes have been achieved in modeling biological neural systems, there are still no solutions to the complex problem of modeling intuition, consciousness and emotion solutions to the complex problem of modeling intuition, consciousness and emotion - - which which form form integral parts of human intelligence”…( integral parts of human intelligence”…(Alan Turing, 1950) ---Human brain has the ability to perform tasks such as pattern recognition, perception and motor control much faster than any computer---
  • 3. Facts of Human Brain (complex, nonlinear and parallel computer) • The brain contains about 1010 (100 billion) basic units called neurons • Each neuron connected to about 104 other neurons • Weight: birth 0.3 kg, adult ~1.5 kg • Power consumption 20-40W (~20% of body consumption) • Signal propagation speed inside the axon ~90m/s in ~170,000 Km of axon length for adult male • Firing frequency of a neuron ~250 – 2000Hz • Operating temperature: 37±2oC • Sleep requirement: average 7.5 hours (adult) Intel Pentium 4 1.5GHz Number of transistors 4.2x107 Power consumption up to 55 Watts Weight 0.1 kg cartridge w/o fans, 0.3 kg with fan/heatsink Maximum firing frequency 1.5 GHz Normal operating temperature 15-85°C Sleep requirement 0 (if not overheated/ overclocked) Processing of complex stimuli if can be done, takes a long time
  • 4. Biological neuron • Soma: Nucleus of neuron (the cell body) - process the input • Dendrites: long irregularly shaped filaments attached to the soma – input channels • Axon: another type link attached to the soma – output channels • Output of the axon: voltage pulse (spike) voltage pulse (spike) that lasts for a ms • Firing of neuron – membrane potential • Axon terminates in a specialized contact called the synaptic junction – the electrochemical contact between neurons • The size of synapses are believed to be linked with learning • Larger area: excitatory—smaller area: inhibitory
  • 5. Artificial neuron model (McCulloh-Pitts model, 1949) Qj : external threshold, offset or bias wji : synaptic weights xi : input yj : output …..Another model-Product unit Firing and the strength of the exiting signal are controlled by activation function (AF) Allow higher-order combinations of inputs, having the advantage of increased information capacity Types of AF: • •Linear Linear • •Step Step • •Ramp Ramp • •Sigmoid Sigmoid • •Hyperbolic tangent Hyperbolic tangent • •Gaussian Gaussian
  • 6. Different NN types • Single-layer NNs, such as the Hopfield network • Multilayer feedforward NNs, for example standard backpropagation, functional link and product unit networks • Temporal NNs, such as the Elman and Jordan simple recurrent networks as well as time-delay neural networks • Self-organizing NNs, such as the Kohonen self-organizing feature maps and the learning vector quantizer • Combined feedforward and self-organizing NNs, such as the radial basis function networks
  • 7. The ANN applications • • Classification, Classification, the aim is to predict the class of an input vector • • Pattern matching Pattern matching, the aim is to produce a pattern best associated with a given input vector • • Pattern completion Pattern completion, the aim is to complete the missing parts of a given input vector • • Optimization Optimization, the aim is to find the optimal values of parameters in an optimization problem • • Control Control, an appropriate action is suggested based on given an input vectors • • Function approximation/times series modeling Function approximation/times series modeling, the aim is to learn the functional relationships between input and desired output vectors; • • Data mining Data mining, with the aim of discovering hidden patterns from data (knowledge discovery)
  • 8. ANN architectures • Neural Networks are known to be universal function approximators • Various architectures are available to approximate any nonlinear function • Different architectures allow for generation of functions of different complexity and power Feedforward networks Feedback networks Lateral networks
  • 9. Feedforward Networks Network size: n x m x r = 2x5x1 Wmn: input weight matrix Vrm: output weight matrix •No feedback within the network •The coupling takes place from one layer to the next •The information flows, in general, in the forward direction Input layer: Number of neurons in this layer corresponds to the number of inputs to the neuronal network. This layer consists of passive nodes, i.e., which do not take part in the actual signal modification, but only transmits the signal to the following layer. • Hidden layer: This layer has arbitrary number of layers with arbitrary number of neurons. The nodes in this layer take part in the signal modification, hence, they are active. • Output layer: The number of neurons in the output layer corresponds to the number of the output values of the neural network. The nodes in this layer are active ones. FFNN can have more than one hidden layer. However, it has been proved that FFNNs with one hidden layer has enough to approximate any continuous function [Hornik 1989].
  • 10. Feedback networks Elman Recurrent Network The output of a neuron is either directly or indirectly fed back to its input via other linked neurons used in complex pattern recognition tasks, e.g., speech recognition etc.
  • 12. Lateral Networks •There exist couplings of neurons within one layer •There is no essentially explicit feedback path amongst the different layers •This can be thought of as a compromise between the forward and feedback network Input layer Hidden layer Output layer
  • 13. Learning methods • Artificial neural networks work through the optimized weight values. • The method by which the optimized weight values are attained is called learning learning • In the learning process  try to teach the network how to produce the output when the corresponding input is presented • When learning is complete: the trained neural network, with the updated optimal weights, should be able to produce the output within desired accuracy corresponding to an input pattern. Learning methods • Supervised learning • Unsupervised learning • Reinforced learning
  • 15. Supervised learning Supervised learning means guided learning by “teacher”; requires a training set which consists of input vectors and a target vector associated with each input vector “Learning experience in our childhood” As a child, we learn about various things (input) when we see them and simultaneously are told (supervised) about their names and the respective functionalities (desired response). Supervised learning system: •feedforward •functional link •product unit •Recurrent •Time delay
  • 16. Unsupervised learning • The objective of unsupervised learning is to discover patterns or features in the input data with no help from a teacher, basically performing a clustering of input space. • The system learns about the pattern from the data itself without a priori knowledge. This is similar to our learning experience in adulthood “For example, often in our working environment we are thrown into a project or situation which we know very little about. However, we try to familiarize with the situation as quickly as possible using our previous experiences, education, willingness and similar other factors” • Hebb’s rule: It helps the neural network or neuron assemblies to remember specific patterns much like the memory. From that stored knowledge, similar sort of incomplete or spatial patterns could be recognized. This is even faster than the delta rule or the backpropagation algorithm because there is no repetitive presentation and training of input–output pairs.
  • 17. Reinforced learning •A ‘teacher’ though available, does not present the expected answer but only indicates if the computed output is correct or incorrect •The information provided helps the network in its learning process •A reward is given for a correct answer computed and a penalty for a wrong answer
  • 18. Leaning algorithm in Supervised learning • Gradient descent • Widrow-hoff (LMS) • Generalized delta • Error-correction Single neuron
  • 19. Gradient Descent • Gradient descent (GD)…(not the first but used most) • GD is aimed to find the weight values that minimize Error • GD requires the definition of an error (or objective) function to measure the neuron's error in approximating the target Analogy: Suppose we want to come down (descend) from a high hill (higher error) to a low valley (lower error). We move along the negative gradient or slopes. By doing so, we take the steepest path to the downhill valley steepest descent algorithm Where tp and fp are respectively the target and actual output for patterns p The updated weights: where η :learning rate wi(t+1):new weights The calculation of the partial derivative of f with respect to up (the net input for pattern p) presents a problem for all discontinuous activation functions, such as the step and ramp functions Widrow-Hoff learning rule
  • 20. Widrow-hoff Least-Means-Square (LMS) Assume that f = up The weights are updated using: One of the first algorithms used to train multiple adaptive linear neurons (Madaline) [Widrow 1987, Widrow and Lehr 1990]
  • 21. Generalized delta Assume: differentiable activation functions; such as sigmoid function The weights are updated using:
  • 22. Error-correction Assume that binary-valued functions are used, e.g the step function. The weights are updated using: Weights are only adjusted when the neuron responds in error
  • 23. Feedforward neural network with Gradient descent optimization Input vectors  actual value is calculated then error is calculated The error gradient respect to network’s weight is calculated by propagating the error backward through network Once the error gradient is calculated, the weight is adjusted More details…
  • 25. Feedforward Operation Input vector xj where j =1 to n (number of inputs) Input weight matrix Wij where i = 1 to m (hidden neurons) Step 1: Activation vector ai : Decision vector di : Step 2: Output vector yi is given by (r is no. of outputs): 
  • 26. Backpropagation Operation Step 1: The output error vector: Step 2: The decision error vector: The activation error vector: Step 3: The weights changes: gg and gm are learning and momentum rates, respectively The weights updates: One set of weight modifications is called an epoch, and many of these may be required before the desired accuracy of approximation is reached. “This is the objective function for NN learning that need to be optimized optimized by the optimization methods” The backpropagation training algorithm is based on the principle of gradient descent and is given as half the square of the Euclidean norm of the output error vector.
  • 27. Optimization methods to carry out NN learning • Local optimization, where the algorithm ends up in a local optimum without finding a global optimum. Gradient descent and scaled conjugate gradient are local optimizers. • Global optimization, where the algorithm searches for the global optimum by with mechanisms that allow greater search space explorations. Global optimizers include Leapfrog, simulated annealing, evolutionary computing and swarm optimization. “Local and global optimization techniques can be combined to form hybrid training algorithms”
  • 28. Weight Adjustments/Updates • Stochastic/Delta/(online) learning, where the NN weights are adjusted after each pattern presentation. In this case the next input pattern is selected randomly from the training set, to prevent any bias that may occur due to the sequences in which patterns occur in the training set. • Batch/(offline) learning, where the NN weight changes are accumulated and used to adjust weights only after all training patterns have been presented Two types of supervised learning algorithms exist, based on when/how weights are updated:
  • 29. Feedforward Neural Networks (effects of weight variations)
  • 30. Homework! • Please make a report about the potential of intelligent techniques is applied for the part of your current research • Due date: 12 May 2010 Project! Critical review Please find very recent paper about the application of intelligent techniques in power system area; try to understand the paper and criticize the contents Due date: will be at the final lecture