7. Hebb’s Rule
• 1949. Increase the weight between two
neurons that are both “on”.
• 1988. Increase the weight between two
neurons that are both “off”.
• wi(new) = wi(old) + xi*y
8. Algorithm
1. set wi = 0 for 0 <= i <= n
2. for each training vector
3. set xi = si for all input units
4. set y = t
5. wi(new) = wi(old) + xi*y
13. Pattern Recognition Exercise
• Architecture?
• Weights?
• Are the original patterns classified
correctly?
• Are the original patterns with 1 piece of
wrong data classified correctly?
• Are the original patterns with 1 piece of
missing data classified correctly?
14. Perceptrons (1958)
• Very important early neural network
• Guaranteed training procedure under
certain circumstances
x0
x1
y
xn
1
w0
w1
wn
15. Activation Function
• f(yin) = 1 if yin >
f(yin) = 0 if - <= yin <=
f(yin) = -1 otherwise
• Graph interpretation
1
-1
16. Learning Rule
• wi(new) = wi(old) + *t*xi if error
• is the learning rate
• Typically, 0 < <= 1
17. Algorithm
1. set wi = 0 for 0 <= i <= n (can be random)
2. for each training exemplar do
3. xi = si
4. yin = xi*wi
5. y = f(yin)
6. wi(new) = wi(old) + *t*xi if error
7. if stopping condition not reached, go to 2
21. Perceptron Learning Rule
Convergence Theorem
• If a weight vector exists that correctly
classifies all of the training examples, then
the perceptron learning rule will converge
to some weight vector that gives the
correct response for all training patterns.
This will happen in a finite number of
steps.
23. Adaline (Widrow, Huff 1960)
• Adaptive Linear Network
• Learning rule minimizes the mean squared
error
• Learns on all examples, not just ones with
errors
25. Training Algorithm
1. set wi (small random values typical)
2. set (0.1 typical)
3. for each training exemplar do
4. xi = si
5. yin = xi*wi
6. wi(new) = wi(old) + *(t – yin)*xi
7. go to 3 if largest weight change big
enough