SlideShare a Scribd company logo
ARTIFICIAL NEURAL NETWORKS
AIMS Education
Introduction
• Simple computational elements forming a large
network
– Emphasis on learning (pattern recognition)
– Local computation (neurons)
• Configured for a particular application
– Pattern recognition/data classification
• ANN algorithm
– Modeled after brain
• Brain: 100,000 times slower response
– Complex tasks (image, sound recognition, motion con)
– –10,000,000,000 times efficient in energy
consumption/op
AIMS Education
Introduction (Contd….)
• Artificial Intelligence
• Structure
– Inputs vs Dendrites
– Weights vs Synaptic gap
– Neurons vs Soma
– Output vs Axon
AIMS Education
Introduction (Contd….)
AIMS Education
History
AIMS Education
Definition
A neural network is a massively parallel
distributed processor made up of simple
processing units, which has a natural tendency
for storing experiential knowledge and making
it available for use
AIMS Education
Introduction (Contd….)
• The threshold value determine the final output
– If the summation < threshold, -1 is the output
– If the summation > threshold, +1 is the output
AIMS Education
Introduction (Contd….)
• The neuron is the basic information
processing unit of an ANN.
• It consists of:
– A set of links, describing the neuron inputs, with
weightsW1, W2, …, Wm
– An adder function (linear combiner) for computing
the weighted sum of the inputs (real numbers):
AIMS Education
Introduction (Contd….)
– Activation function(squashing function) for
limiting the amplitude of the neuron output.
• The bias b has the effect of applying an affine
transformation to the weighted sum u
v = u + b
AIMS Education
Introduction (Contd….)
AIMS Education
AIMS Education
Activation Functions
AIMS Education
Designing an ANN
• Designing an ANN consist of
– Arranging neurons in various layers
– Deciding the type of connections among neurons for
different layers, as well as among the neurons within a
layer
– Deciding the way a neuron receives input and
produces output
• Determining the strength of connection within
the network by allowing the network learn the
appropriate values of connection weights by
using a training data set.
• The process of designing a neural network is an
iterative process
AIMS Education
Designing an ANN (Contd….)
• Layers
– Single layer: Input and output layer
• No computation at the input layer
• Computation takes place at the output layer
– Multi layer: Input, hidden(s) and output layers
• Computations are done at the hidden(s) and the output layer
• Connection
– Fully connected
• Each neuron on the first layer is connected to every neuron on
the second layer
– Partially connected
• A neuron of the first layer does not have to be connected to all
neurons on the second layer
AIMS Education
Designing an ANN (Contd….)
• Very complex webs of interconnected neurons
– Simple interconnected units in ANNs
• Each unit takes in a number of real-valued
inputs
– Possibly outputs of other units
• Produces a single real-valued output
– May become the input to many other units
AIMS Education
Single Layer
AIMS Education
Multi Layer
AIMS Education
Appropriate problems for Neural
Network Learning
• Instances are represented by many attribute-
value pairs
– The target function to be learned is defined over
instances that can be described by a vector of
predefined features
– Input attributes may be highly correlated or
independent of one another
– Input values can be any real values
AIMS Education
Appropriate problems for Neural
Network Learning (Contd….)
• The target function output may be discrete-
valued, real-valued, or a vector of several real- or
discrete-valued attributes
• Training examples may contain errors
– Robust to noisy data
• Long training times are acceptable
– Network training algorithms typically require longer
training times
– Depends on factors such as
• The number of weights in the network
• The number of training examples considered
AIMS Education
Appropriate problems for Neural
Network Learning (Contd….)
• Fast evaluation of the learned target function
may be required
– Learning times are relatively long, evaluating the
learned network, in order to apply it to a
subsequent instances, is typically very fast
• The ability of humans to understand the
learned target function is not important
– Weights learned are often difficult for humans to
interpret
AIMS Education
Perceptron
AIMS Education
Perceptron (Contd….)
• Takes a vector of real-valued inputs
• Calculates a linear combination of these
inputs
• Outputs a 1 if the result is greater than some
threshold
• Outputs a -1 otherwise
• Weights: An information that allows ANN to
achieve the desired results. This information
changes during the learning process
AIMS Education
Perceptron (Contd….)
• Given inputs 𝑥1 through 𝑥𝑛, the output 𝑜(𝑥
1,…,n𝑛) computed by the perceptron is
𝑜(𝑥1,…,𝑥𝑛)= 1 𝑖𝑓
𝑤0+ 𝑤1𝑥1+ 𝑤2𝑥2+ …+ 𝑤𝑛𝑥𝑛 > 0
−1 other𝑤𝑖𝑠𝑒
• Where each 𝑤𝑖 is a real-valued constant, or
weight, that determines the contribution of
input 𝑥𝑖 to the perceptron output
AIMS Education
Perceptron (Contd….)
• An additional constant input 𝑥0=1, allowing us
to write the inequality as
∑i=1 to n 𝑤𝑖𝑥𝑖>0
In vector form as 𝑤.𝑥 > 0
• For brevity, we will sometimes write the
perceptron function as 𝑜(𝑥) =𝑠𝑔𝑛(𝑤.𝑥) where
𝑠𝑔𝑛(𝑦)= 1 𝑖𝑓 𝑦 > 0
-1 otherwise
AIMS Education
Representational Power of
Perceptrons
• A perceptron can be seen as representing a
hyper plane decision surface in the 𝑛-
dimensional space of instances (i.e., points)
• It outputs a 1 for instances lying on one side of
the hyper place
• Outputs a −1 for instances lying on the other
side
AIMS Education
Representational Power of
Perceptrons (Contd….)
AIMS Education
Representational Power of
Perceptrons (Contd….)
• The equation for the decision surface is 𝑤.𝑥 =0
• Some sets of positive and negative examples
cannot be separated by any hyperplace
• The ones that can be separated are called
linearly separable sets of examples
AIMS Education
Perceptron Training Rule
• How to learn the weights for a single perceptron?
• The task is to learn a weight vector that causes
the perceptron to produce the correct ±1 output
for each of the given training examples
• Perceptron rule and delta rule algorithms
– Provide the basis for learning networks of many units
AIMS Education
Perceptron Training Rule (Contd….)
• One way to learn an acceptable weight vector is to
– Begin with random weights
– Iteratively apply the perceptron to each training
example
– Modify the perceptron weights whenever it
misclassifies an example
– The process is repeated
• Iterating through the training examples as many times
as needed
• Until the perceptron classifies all training examples
correctly
AIMS Education
Perceptron Training Rule (Contd….)
• Weights are modified at each step according to the
perceptron training rule
• Here 𝑡 is the target output for the current training example
• 𝑜 is the output generated by the perceptron
• 𝜂 is a positive constant called the learning rate
– Moderates the degree to which weights are changed at each
step
– Usually set to some small value (e.g., 0.1)
– Sometimes made to decay as the number of weight-tuning
iterations increases
AIMS Education
Perceptron Training Rule (Contd….)
• Why should it converge to successful weight
values?
– Suppose the training example is correctly classified
already by the perceptron
– In this case 𝑡 −𝑜 is zero
– Makes Δ𝑤𝑖 zero
– No weights are updated
– Suppose it outputs a -1 when the target output is +1
– Weights must be altered to increase the value of (𝑤.𝑥)
• For example, if 𝑥𝑖>0, then increasing 𝑤𝑖 will bring the
perceptron closer to correctly classifying in this example
• Can be shown to converge within a finite number of
applications of the perceptron training rule
AIMS Education
Gradient Descent and the Delta Rule
• Perceptron rule works fine when the training
examples are linearly separable
– Otherwise can fail to converge
• Delta rule is defined to overcome this hurdle
• If training examples are not linearly separable
• Delta rule converges to the best fit
approximation to the target concept
AIMS Education
Delta Rule (Contd….)
• Becomes the basis for learning interconnected
networks (multilayer network)
• Training an unthresholded perceptron, a linear
unit for which the output 𝑜 is given by
– 𝑜(𝑥) = (w.x)
– It corresponds to the first stage of a perceptron,
without the threshold
AIMS Education
Delta Rule (Contd….
• Training error (weight vector), relative to the
training examples
• Where 𝐷 is the set of training examples
• 𝑡𝑑 is the target output for the training example 𝑑
• 𝑜𝑑 is the output of the linear unit for training
example 𝑑
• is simply half the squared difference
between the target output 𝑡𝑑 and the linear unit
output 𝑜𝑑 summed over all training examples
AIMS Education

More Related Content

PPTX
Convolution Neural Network (CNN)
PPTX
Types of Machine Learning
PPTX
Deep learning
PPTX
lazy learners and other classication methods
PDF
I. AO* SEARCH ALGORITHM
PPTX
MACHINE LEARNING - GENETIC ALGORITHM
PPT
Deep Learning
PDF
I.INFORMED SEARCH IN ARTIFICIAL INTELLIGENCE II. HEURISTIC FUNCTION IN AI III...
Convolution Neural Network (CNN)
Types of Machine Learning
Deep learning
lazy learners and other classication methods
I. AO* SEARCH ALGORITHM
MACHINE LEARNING - GENETIC ALGORITHM
Deep Learning
I.INFORMED SEARCH IN ARTIFICIAL INTELLIGENCE II. HEURISTIC FUNCTION IN AI III...

What's hot (20)

PPTX
Neural networks
PPT
Iterative deepening search
PDF
I. Hill climbing algorithm II. Steepest hill climbing algorithm
PPTX
Graph coloring using backtracking
PPTX
Inductive analytical approaches to learning
PDF
Anomaly Detection using Deep Auto-Encoders
PPTX
Artifical Neural Network and its applications
PPT
Backtracking
PPTX
Flowchart of GA
PPTX
Supervised learning and Unsupervised learning
PPTX
Using prior knowledge to initialize the hypothesis,kbann
PDF
State Space Search in ai
PPTX
Image classification using convolutional neural network
PDF
I. Mini-Max Algorithm in AI
PPTX
8 queens problem using back tracking
PPTX
Problem solving agents
ODP
Genetic algorithm ppt
PPTX
Genetic Algorithm in Artificial Intelligence
PPTX
Batch normalization presentation
PPTX
Supervised and unsupervised learning
Neural networks
Iterative deepening search
I. Hill climbing algorithm II. Steepest hill climbing algorithm
Graph coloring using backtracking
Inductive analytical approaches to learning
Anomaly Detection using Deep Auto-Encoders
Artifical Neural Network and its applications
Backtracking
Flowchart of GA
Supervised learning and Unsupervised learning
Using prior knowledge to initialize the hypothesis,kbann
State Space Search in ai
Image classification using convolutional neural network
I. Mini-Max Algorithm in AI
8 queens problem using back tracking
Problem solving agents
Genetic algorithm ppt
Genetic Algorithm in Artificial Intelligence
Batch normalization presentation
Supervised and unsupervised learning
Ad

Similar to ARTIFICIAL NEURAL NETWORKS (20)

PDF
Machine Learning- Perceptron_Backpropogation_Module 3.pdf
PPT
Neural
PPS
Neural Networks
PDF
Artificial Neural Network
PPT
PDF
Artificial Neural Network
PPT
SOFT COMPUTERING TECHNICS -Unit 1
PDF
m3 (2).pdf
PPTX
Chapter-5-Part I-Basics-Neural-Networks.pptx
PPT
2011 0480.neural-networks
PDF
071bct537 lab4
PPTX
Artificial Neural Network by Dr.C.R.Dhivyaa Kongu Engineering College
PPTX
UNIT III (8).pptx
PPTX
Neural network
PDF
ARTIFICIAL-NEURAL-NETWORKMACHINELEARNING
PPT
The Introduction to Neural Networks.ppt
PPT
Artificial-Neural-Networks.ppt
PDF
Deep Learning detailkesdSECA4002 doc.pdf
PDF
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
PPTX
Artificial Neural Networks for NIU session 2016 17
Machine Learning- Perceptron_Backpropogation_Module 3.pdf
Neural
Neural Networks
Artificial Neural Network
Artificial Neural Network
SOFT COMPUTERING TECHNICS -Unit 1
m3 (2).pdf
Chapter-5-Part I-Basics-Neural-Networks.pptx
2011 0480.neural-networks
071bct537 lab4
Artificial Neural Network by Dr.C.R.Dhivyaa Kongu Engineering College
UNIT III (8).pptx
Neural network
ARTIFICIAL-NEURAL-NETWORKMACHINELEARNING
The Introduction to Neural Networks.ppt
Artificial-Neural-Networks.ppt
Deep Learning detailkesdSECA4002 doc.pdf
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
Artificial Neural Networks for NIU session 2016 17
Ad

More from AIMS Education (20)

PDF
The Next Chapter (The Evolving Relationship between America and Pakistan )
PDF
InterContinental Hotels Group (IHG) Annual Report and Form 20-F 2016
DOCX
FACTORS AFFECTING THE BOND MARKET
DOCX
A SURVEY REPORT ON TEVTA PAKISTAN
PPTX
A Quick Tour of Logos: The Logical Appeal
PPTX
Parafait – POS Training Manual
PPTX
Social Media - Evolution And Revolution - Start to Today
PPTX
How LinkedIn built a Community of Half a Billionaire
DOCX
Format of Appointment Letter
DOCX
Format of Bank Account Opening Request
PPTX
Selection of ERP vendor
PPTX
ERP Lecture 3&4 (practical screen shots)
PPTX
REFRENCING
PPTX
Quantitative Research
PPTX
Qualitative Research
PPTX
Research Design and Proposal Writing
PPTX
TRAINING AND SELECTION
PPTX
RECRUITMENT AND SELECTION
PPTX
HUMAN RESOURCE PLANNING
PPTX
HUMAN RESOURCE
The Next Chapter (The Evolving Relationship between America and Pakistan )
InterContinental Hotels Group (IHG) Annual Report and Form 20-F 2016
FACTORS AFFECTING THE BOND MARKET
A SURVEY REPORT ON TEVTA PAKISTAN
A Quick Tour of Logos: The Logical Appeal
Parafait – POS Training Manual
Social Media - Evolution And Revolution - Start to Today
How LinkedIn built a Community of Half a Billionaire
Format of Appointment Letter
Format of Bank Account Opening Request
Selection of ERP vendor
ERP Lecture 3&4 (practical screen shots)
REFRENCING
Quantitative Research
Qualitative Research
Research Design and Proposal Writing
TRAINING AND SELECTION
RECRUITMENT AND SELECTION
HUMAN RESOURCE PLANNING
HUMAN RESOURCE

Recently uploaded (20)

PPTX
Entre CHtzyshshshshshshshzhhzzhhz 4MSt.pptx
PPTX
sdn_based_controller_for_mobile_network_traffic_management1.pptx
PPT
FABRICATION OF MOS FET BJT DEVICES IN NANOMETER
PPTX
STEEL- intro-1.pptxhejwjenwnwnenemwmwmwm
PPTX
making presentation that do no stick.pptx
PPTX
02fdgfhfhfhghghhhhhhhhhhhhhhhhhhhhh.pptx
PPTX
5. MEASURE OF INTERIOR AND EXTERIOR- MATATAG CURRICULUM.pptx
PDF
How NGOs Save Costs with Affordable IT Rentals
PPTX
Embedded for Artificial Intelligence 1.pptx
PPT
Hypersensitivity Namisha1111111111-WPS.ppt
PDF
Smarter Security: How Door Access Control Works with Alarms & CCTV
PPTX
Lecture 3b C Library _ ESP32.pptxjfjfjffkkfkfk
PPTX
Prograce_Present.....ggation_Simple.pptx
PPTX
Embeded System for Artificial intelligence 2.pptx
PPTX
material for studying about lift elevators escalation
PPTX
DEATH AUDIT MAY 2025.pptxurjrjejektjtjyjjy
PDF
Cableado de Controladores Logicos Programables
PPTX
Syllabus Computer Six class curriculum s
PPTX
Lecture-3-Computer-programming for BS InfoTech
PPTX
1.pptxsadafqefeqfeqfeffeqfqeqfeqefqfeqfqeffqe
Entre CHtzyshshshshshshshzhhzzhhz 4MSt.pptx
sdn_based_controller_for_mobile_network_traffic_management1.pptx
FABRICATION OF MOS FET BJT DEVICES IN NANOMETER
STEEL- intro-1.pptxhejwjenwnwnenemwmwmwm
making presentation that do no stick.pptx
02fdgfhfhfhghghhhhhhhhhhhhhhhhhhhhh.pptx
5. MEASURE OF INTERIOR AND EXTERIOR- MATATAG CURRICULUM.pptx
How NGOs Save Costs with Affordable IT Rentals
Embedded for Artificial Intelligence 1.pptx
Hypersensitivity Namisha1111111111-WPS.ppt
Smarter Security: How Door Access Control Works with Alarms & CCTV
Lecture 3b C Library _ ESP32.pptxjfjfjffkkfkfk
Prograce_Present.....ggation_Simple.pptx
Embeded System for Artificial intelligence 2.pptx
material for studying about lift elevators escalation
DEATH AUDIT MAY 2025.pptxurjrjejektjtjyjjy
Cableado de Controladores Logicos Programables
Syllabus Computer Six class curriculum s
Lecture-3-Computer-programming for BS InfoTech
1.pptxsadafqefeqfeqfeffeqfqeqfeqefqfeqfqeffqe

ARTIFICIAL NEURAL NETWORKS

  • 2. Introduction • Simple computational elements forming a large network – Emphasis on learning (pattern recognition) – Local computation (neurons) • Configured for a particular application – Pattern recognition/data classification • ANN algorithm – Modeled after brain • Brain: 100,000 times slower response – Complex tasks (image, sound recognition, motion con) – –10,000,000,000 times efficient in energy consumption/op AIMS Education
  • 3. Introduction (Contd….) • Artificial Intelligence • Structure – Inputs vs Dendrites – Weights vs Synaptic gap – Neurons vs Soma – Output vs Axon AIMS Education
  • 7. Definition A neural network is a massively parallel distributed processor made up of simple processing units, which has a natural tendency for storing experiential knowledge and making it available for use AIMS Education
  • 8. Introduction (Contd….) • The threshold value determine the final output – If the summation < threshold, -1 is the output – If the summation > threshold, +1 is the output AIMS Education
  • 9. Introduction (Contd….) • The neuron is the basic information processing unit of an ANN. • It consists of: – A set of links, describing the neuron inputs, with weightsW1, W2, …, Wm – An adder function (linear combiner) for computing the weighted sum of the inputs (real numbers): AIMS Education
  • 10. Introduction (Contd….) – Activation function(squashing function) for limiting the amplitude of the neuron output. • The bias b has the effect of applying an affine transformation to the weighted sum u v = u + b AIMS Education
  • 14. Designing an ANN • Designing an ANN consist of – Arranging neurons in various layers – Deciding the type of connections among neurons for different layers, as well as among the neurons within a layer – Deciding the way a neuron receives input and produces output • Determining the strength of connection within the network by allowing the network learn the appropriate values of connection weights by using a training data set. • The process of designing a neural network is an iterative process AIMS Education
  • 15. Designing an ANN (Contd….) • Layers – Single layer: Input and output layer • No computation at the input layer • Computation takes place at the output layer – Multi layer: Input, hidden(s) and output layers • Computations are done at the hidden(s) and the output layer • Connection – Fully connected • Each neuron on the first layer is connected to every neuron on the second layer – Partially connected • A neuron of the first layer does not have to be connected to all neurons on the second layer AIMS Education
  • 16. Designing an ANN (Contd….) • Very complex webs of interconnected neurons – Simple interconnected units in ANNs • Each unit takes in a number of real-valued inputs – Possibly outputs of other units • Produces a single real-valued output – May become the input to many other units AIMS Education
  • 19. Appropriate problems for Neural Network Learning • Instances are represented by many attribute- value pairs – The target function to be learned is defined over instances that can be described by a vector of predefined features – Input attributes may be highly correlated or independent of one another – Input values can be any real values AIMS Education
  • 20. Appropriate problems for Neural Network Learning (Contd….) • The target function output may be discrete- valued, real-valued, or a vector of several real- or discrete-valued attributes • Training examples may contain errors – Robust to noisy data • Long training times are acceptable – Network training algorithms typically require longer training times – Depends on factors such as • The number of weights in the network • The number of training examples considered AIMS Education
  • 21. Appropriate problems for Neural Network Learning (Contd….) • Fast evaluation of the learned target function may be required – Learning times are relatively long, evaluating the learned network, in order to apply it to a subsequent instances, is typically very fast • The ability of humans to understand the learned target function is not important – Weights learned are often difficult for humans to interpret AIMS Education
  • 23. Perceptron (Contd….) • Takes a vector of real-valued inputs • Calculates a linear combination of these inputs • Outputs a 1 if the result is greater than some threshold • Outputs a -1 otherwise • Weights: An information that allows ANN to achieve the desired results. This information changes during the learning process AIMS Education
  • 24. Perceptron (Contd….) • Given inputs 𝑥1 through 𝑥𝑛, the output 𝑜(𝑥 1,…,n𝑛) computed by the perceptron is 𝑜(𝑥1,…,𝑥𝑛)= 1 𝑖𝑓 𝑤0+ 𝑤1𝑥1+ 𝑤2𝑥2+ …+ 𝑤𝑛𝑥𝑛 > 0 −1 other𝑤𝑖𝑠𝑒 • Where each 𝑤𝑖 is a real-valued constant, or weight, that determines the contribution of input 𝑥𝑖 to the perceptron output AIMS Education
  • 25. Perceptron (Contd….) • An additional constant input 𝑥0=1, allowing us to write the inequality as ∑i=1 to n 𝑤𝑖𝑥𝑖>0 In vector form as 𝑤.𝑥 > 0 • For brevity, we will sometimes write the perceptron function as 𝑜(𝑥) =𝑠𝑔𝑛(𝑤.𝑥) where 𝑠𝑔𝑛(𝑦)= 1 𝑖𝑓 𝑦 > 0 -1 otherwise AIMS Education
  • 26. Representational Power of Perceptrons • A perceptron can be seen as representing a hyper plane decision surface in the 𝑛- dimensional space of instances (i.e., points) • It outputs a 1 for instances lying on one side of the hyper place • Outputs a −1 for instances lying on the other side AIMS Education
  • 27. Representational Power of Perceptrons (Contd….) AIMS Education
  • 28. Representational Power of Perceptrons (Contd….) • The equation for the decision surface is 𝑤.𝑥 =0 • Some sets of positive and negative examples cannot be separated by any hyperplace • The ones that can be separated are called linearly separable sets of examples AIMS Education
  • 29. Perceptron Training Rule • How to learn the weights for a single perceptron? • The task is to learn a weight vector that causes the perceptron to produce the correct ±1 output for each of the given training examples • Perceptron rule and delta rule algorithms – Provide the basis for learning networks of many units AIMS Education
  • 30. Perceptron Training Rule (Contd….) • One way to learn an acceptable weight vector is to – Begin with random weights – Iteratively apply the perceptron to each training example – Modify the perceptron weights whenever it misclassifies an example – The process is repeated • Iterating through the training examples as many times as needed • Until the perceptron classifies all training examples correctly AIMS Education
  • 31. Perceptron Training Rule (Contd….) • Weights are modified at each step according to the perceptron training rule • Here 𝑡 is the target output for the current training example • 𝑜 is the output generated by the perceptron • 𝜂 is a positive constant called the learning rate – Moderates the degree to which weights are changed at each step – Usually set to some small value (e.g., 0.1) – Sometimes made to decay as the number of weight-tuning iterations increases AIMS Education
  • 32. Perceptron Training Rule (Contd….) • Why should it converge to successful weight values? – Suppose the training example is correctly classified already by the perceptron – In this case 𝑡 −𝑜 is zero – Makes Δ𝑤𝑖 zero – No weights are updated – Suppose it outputs a -1 when the target output is +1 – Weights must be altered to increase the value of (𝑤.𝑥) • For example, if 𝑥𝑖>0, then increasing 𝑤𝑖 will bring the perceptron closer to correctly classifying in this example • Can be shown to converge within a finite number of applications of the perceptron training rule AIMS Education
  • 33. Gradient Descent and the Delta Rule • Perceptron rule works fine when the training examples are linearly separable – Otherwise can fail to converge • Delta rule is defined to overcome this hurdle • If training examples are not linearly separable • Delta rule converges to the best fit approximation to the target concept AIMS Education
  • 34. Delta Rule (Contd….) • Becomes the basis for learning interconnected networks (multilayer network) • Training an unthresholded perceptron, a linear unit for which the output 𝑜 is given by – 𝑜(𝑥) = (w.x) – It corresponds to the first stage of a perceptron, without the threshold AIMS Education
  • 35. Delta Rule (Contd…. • Training error (weight vector), relative to the training examples • Where 𝐷 is the set of training examples • 𝑡𝑑 is the target output for the training example 𝑑 • 𝑜𝑑 is the output of the linear unit for training example 𝑑 • is simply half the squared difference between the target output 𝑡𝑑 and the linear unit output 𝑜𝑑 summed over all training examples AIMS Education