NN LMS DR Gamal PDF
NN LMS DR Gamal PDF
NN LMS DR Gamal PDF
CHAPTER 03
THE LEAST-MEAN SQUARE
ALGORITHM
Prof. Dr. Mostafa Gadal-Haqq M. Mostafa
Computer Science Department
Faculty of Computer & Information Sciences
AIN SHAMS UNIVERSITY
Newton’s Method
Gauss-Newton Method
Computer Experiment
Figure 3.2 Trajectory of the method of steepest descent in a two-dimensional space for two
different values of learning-rate parameter: (a) small η (b) large η .The coordinates w1 and w2
are elements of the weight vector w; they both lie in the W -plane.
ASU-CSC445: Neural Networks Prof. Dr. Mostafa Gadal-Haqq 9
Unconstrained Optimization: A Review
Newton’s Method
The idea is to minimize the quadratic approximation of the
cost function E (w) around the current point w(n):
E (w (n)) E (w (n 1)) E (w (n))
1
g T (n)w (n) wT (n)H(n)w (n)
2
Where H(n) is the Hessian matrix of E (w)
H (n) 2E (w (n))
The weights are updated by 2E 2E 2E
minimizing E (w) as: w1
2 w1w2 w1wM
2E 2E 2E
w w
w (n 1) w (n) w (n) 2 1 w22 w2wM
w (n) H 1(n)g(n) E E E
2 2 2
wM w1 wM w2 wM 2
Newton’s Method
16
Cost Function
J(0,1)
1
0
21
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
22
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
23
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
24
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
25
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
26
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
27
The corresponding solution (Contour of the cost function
Figure copyright of Andrew Ng.
28
Virtues and Limitation of the LMS Alg.
Robustness
Since the LMS is model independent, therefore it is robust with
respect to disturbance.
That is, the LMS is optimal in accordance with H norm (the
response of the transfer function T of the estimator).
The philosophy of the optimality is to accommodate with the worst-case
scenario:
• “If you don’t know what you are up against, plan for the worst scenario and
optimize.”
Figure 3.8 Formulation of the optimal H∞ estimation problem. The generic estimation error at the transfer
operator’s output could be the weight-error vector, the explanational error, etc.
Figure 3.7 LMS classification with distance –4, based on the double-moon
configuration of Fig. 1.8.
ASU-CSC445: Neural Networks Prof. Dr. Mostafa Gadal-Haqq 33
Homework 3
•Problems:
•3.3
•Computer Experiment
•3.10, 3.13
34
Next Time
Multilayer Perceptrons
35