Lecture 01-Introduction
Lecture 01-Introduction
E.M
Artificial Neural Networks
2
Definition of ANN
3
History
1943: McCulloch and Pitts model neural
networks based on their understanding of
neurology.
1950s: Farley and Clark
» IBM group that tries to model biological behavior
Perceptron (Rosenblatt 1958)
Association units A1, A2, … extract features
from user input
Output is weighted and associated
Function fires if weighted sum of input exceeds
a threshold.
History
Back-propagation learning method (Werbos
1974)
Three layers of neurons
Input, Output, Hidden
Better learning rule for generic three layer networks
Regenerates interest in the 1980s
Successful applications in medicine, marketing,
risk management, … (1990)
In need for another breakthrough.
Neural Networks
A neural network is a massively parallel,
distributed processor
made up of simple processing units (artificial
neurons).
It resembles the brain in two respects:
– Knowledge is acquired by the network
from its
environment through a learning process
– Synaptic connection strengths among
neurons are used to
store the acquired knowledge.
Neural Networks
We are born with about 100 billion neurons
A neuron may connect to as many as
100,000 other neurons
Signals “move” via electrochemical signals
The synapses release a chemical
transmitter – the sum of which can cause a
threshold to be reached – causing the
neuron to “fire”
Synapses can be inhibitory or excitatory
Inspiration from Neurobiology
8
The Structure of Neurons
10
Each neuron has a threshold value
Each neuron has weighted inputs from
other neurons
The input signals form a weighted sum
If the activation level exceeds the
threshold, the neuron “fires”
11
An Artificial Neuron
Each hidden or output neuron has weighted
input connections from each of the units in
the preceding layer.
The unit performs a weighted sum of its
inputs, and subtracts its threshold value, to
give its activation level.
Activation level is passed through a sigmoid
activation function to determine output.
12
Biological Neural Networks
A biological neuron has three types of main
components; dendrites, soma (or cell body)
and axon.
Dendrites receives signals from other
neurons.
The soma, sums the incoming signals.
When sufficient input is received, the cell
fires; that is it transmit a signal over its axon
to other cells.
• From experience: examples / training data
• Strength of connection between the neurons is stored as a weight-value for the
specific connection.
• Learning the solution to a problem = changing the connection weights
An artificial neuron
14
A neural net consists of a large number of simple
processing elements called
neurons, units, cells or nodes.
Each neuron is connected to other neurons by
means of directed communication links, each with
associated weight.
The weight represent information being used by
the net to solve a problem.
15
Neuron Model
Activation Functions
18
• Neural networks are configured for a specific
application, such as pattern recognition or data
classification, through a learning process
• In a biological system, learning involves
adjustments to the synaptic connections between
neurons
same for artificial neural networks (ANNs)
19
Artificial Neural Network
Synapse
x1 w1
y
∑ f Axon
x2 w2
Dendrite
Activation Function:
yin = x1w1 + x2w2 (y-in) = 1 if y-in >=
and (y-in) = 0
20
A neuron receives input, determines the
strength or the weight of the input, calculates
the total
weighted input, and compares the total
weighted with a value (threshold)
The value is in the range of 0 and 1
If the total weighted input greater than or
equal the threshold value, the neuron will
produce the
output, and if the total weighted input less
than the threshold value, no output will be
produced
Types of networks
Single Layer Feed-forward NN
Input layer projecting into the output layer
x1 w11
w12 ym
w21
yn
x2
w22
output layer
Input layer
23
Multilayer feed-forward NN
One or more hidden layers.
– Input projects only from previous layers onto a
layer.
Typically, only from one layer to the next
2-layer or
1-hidden layer
fully connected network
w11
x1 w11
w13 w12 w12 y1
w21 w21
x2 w22
w23 w22
w31 y2
w31 w32
x3 w33 w32
24
Recurrent NN
– A network with feedback, where some of its inputs are
connected to some of its outputs (discrete time).
Input Outputs
Hidden
nodes
25
Types of Layers
The input layer: Introduces input values into the
network
No activation function or other processing
The hidden layer(s): Perform classification of
features
Two hidden layers are sufficient to solve
any problem
Features imply more layers may be better
The output layer: Functionally just like the hidden
layers
Outputs are passed on to the world
outside the neural network.
Early ANN Models:
Perceptron, ADALINE, Hopfield
Network
Current Models:
Deep Learning Architectures
Multilayer feedforward networks
(Multilayer perceptrons)
Radial Basis Function networks
Self Organizing Networks
Example
x1 w1= 0.5
∑ f y
x2 w2 = 0.3
Activation Function:
yin = x1w1 + x2w2 Binary Step Function
= 0.5,
28
Some Neurons have a Bias
Bias
An artificial neuron:
- computes the weighted sum of its input (called its net input)
- adds its bias
- passes this value through an activation function
We say that the neuron “fires” (i.e. becomes active) if its
output is
above zero.
Applications of ANNs
• Signal processing
• Pattern recognition, e.g. handwritten characters or
face identification.
• Diagnosis or mapping symptoms to a medical
case.
• Speech recognition
• Human Emotion Detection
• Educational Loan Forecasting
31
ANN Design
Number of layers
Apparently, three layers is almost always good
enough and better than four layers.
Also: fewer layers are faster in execution and training
How many hidden nodes?
Many hidden nodes allow to learn more complicated
patterns
Because of overtraining, almost always best to set
the number of hidden nodes too low and then
increase their numbers.
Neural Network Architectures
Even for a basic Neural Network, there are
many design decisions to make:
1. # of hidden layers (depth)
2. # of units per hidden layer (width)
3. Type of activation function (nonlinearity)
4. Form of objective function
33