3 - ANN Part One PDF
3 - ANN Part One PDF
3 - ANN Part One PDF
Networks
part one
Artificial Intelligence for
Control and Identification
Dr. Wilbert G. Aguilar
Mster AR
Outline
2. Model of a neuron
determinism
Human Brain
fuzzy behaviour
sequence of instructions
parallelism
high speed
slow speed
repetitive tasks
programming
uniqueness of solutions
ex. matrix product
adaptation to situations
learning
different solutions
ex. face recognition
if the racket is in the right side and the ball comes from right to left
then move a step backward and cross the racket to left side
if racket is in the right side and the ball comes from left to right then
move a step forward
if racket is
A similar mechanism is
used for speech recognition
or motion control
Outline
2. Model of a neuron
2. Model of a neuron
Biological inspiration
soma
nucleus
axon
synapse
dendrites
2. Model of a neuron
Biological inspiration
2. Model of a neuron
Biological inspiration
milliseconds
Number of neurons
1011 1012
Number of connections
Number of synapses
Distributed control
any CPU
Fault tolerant
graceful degradation
10
2. Model of a neuron
McCulloch & Pitts Model (1943)
11
2. Model of a neuron
McCulloch & Pitts Model (1943)
synaptic
weights
x1
w1
x2
xm
xi{-1,1}
wi{-1,1}
y{-1,1}
summing
junction
w2
activation
function
output
wm
1
1 if
0
-1
-2
y=
-1
+1
Sign Function
+2
w x
i
i=1
-1 otherwise
12
2. Model of a neuron
synaptic
weights
x1
w1
summing
junction
x2
activation
function
output
w2
13
2. Model of a neuron
McCulloch & Pitts Model (1943)
Some interesting points:
2. Model of a neuron
Hebbs Rule (1949)
Hebb, D.O. The organization of behavior; a neuropsychological theory. WileyInterscience, New York, 1949
2. Model of a neuron
Hebbs Rule (1949)
w ji = xi y j
16
2. Model of a neuron
The Perceptron (1958)
Rosenblatt, Frank (1958), The Perceptron: A Probabilistic Model for Information Storage and
Organization in the Brain, Cornell Aeronautical Lab, Psychological Review, v65, N.6, pp.386-408.
x1
w1
x2
w2
summing
junction
xm
wm
inputs
synaptic
weights
y
output
activation
function
17
2. Model of a neuron
The Perceptron (1958)
It produces associations between inputs and outputs:
\ {0,1}
Artificial Neural Networks
Pk Tk
k = 1... p
{Pk , Tk }k =1... p
18
2. Model of a neuron
The Perceptron (1958)
w ji = ( Tkj y j ) Pk i
learning rate
desired output
actual output
w ji (t + 1) = w ji (t ) + ( Tkj y j (t ) ) Pk i
i = 1...m
j = 1...n
k = 1... p
19
2. Model of a neuron
The Perceptron (1958)
Novikoff, A. B. (1962). On convergence proofs on perceptrons. Symposium on the
Mathematical Theory of Automata, 12, 615-622. Polytechnic Institute of Brooklyn.
for
Pk
inputs,
t < | y j (t ) = Tkj
20
2. Model of a neuron
The Perceptron (1958)
Vector notation:
x0
w0
x1
w1
x2
w2
xm
wm
x = [x0 x1 x2 ... xm]T
w = [w0 w1 w2 ... wm]T
y = F(wTx) = F(wxT)
21
2. Model of a neuron
x2
C
x1
w2 =
w1
w2
x1
w0
w2
w0=-1
w1=-1 A
w2=1 w =1
0
w1=-1
w2=1
B
w0=1
w1=1
w2=1
w0=-1
w1=1
w2=1
22
2. Model of a neuron
The Perceptron (1958)
Example
(0,1)
(1,1)
(-1,1)
(1,0)
(-1,0)
(0,-1)
\2
{0,1}
(-1,1)
(-1,0)
(0,-1)
(1,0)
(0,1)
(1,1)
23
2. Model of a neuron
The Perceptron (1958)
Minsky M L and Papert S A 1969 Perceptrons (Cambridge, MA: MIT Press)
(0,1)
(0,0)
(1,1)
(1,0)
\2
{0,1}
(0,0)
(0,1)
(1,0)
(1,1)
24
2. Model of a neuron
The Perceptron (1958)
The XOR problem Minsky & Papert (1969)
(0,1)
(0,0)
(1,1)
(1,0)
\2
{0,1}
(0,0)
(0,1)
(1,0)
(1,1)
2. Model of a neuron
The Perceptron (1958)
Synthesis :
Architecture
Transfer function
Hardlim
Associations
\ m {0,1}
Learning rule
w ji (t + 1) = w ji (t ) + ( Tkj y j (t ) ) Pk i
i = 1...m
j = 1...n
k = 1... p
ERROR
26
2. Model of a neuron
Learning
The aim of the learning is to find an association between the
patterns and the targets of the training set
actual
output
p(1,1)
P= #
p(p,1)
y(1,1)
Y= #
y(p,1)
" p(1,m)
%
"
" y(1,n)
%
"
p(p,m)
t(1,1)
T= #
t(p,1)
y(p,n)
t(1,1) - y(1,1)
E=T-Y= #
t(p,1) - y(p,1)
desired
output
" t(1,n)
%
"
t(p,n)
= [ 0]
t(p,n) - y(p,n)
27
2. Model of a neuron
Learning
so the aim of the learning is to find the weights that
minimizes the error, but
Objective function
p
1 p n 2
e = e(w ) = E ( k,j) = diag ( ET E )
2 k =1 j=1
k =1
LEAST
SQUARE
ERROR
28
2. Model of a neuron
Learning
2
1 p n 2
1 p n
min e = e(w ) = E ( k,j) = ( T ( k,j) Y ( k,j) )
2 k =1 j=1
2 k =1 j=1
where
error
surface
29