Professional Documents
Culture Documents
09 Artificial Neural Networks and Classification
09 Artificial Neural Networks and Classification
09 Artificial Neural Networks and Classification
Classification
An artificial neural network is a simple
brain-like device that can learn by
adjusting connections between its
neurons
Processing involves
The neuron
a centimetre or longer.
A real neuron
Signal propagation
Storage capacity
Speed
Designing a network
Network topology
Transfer function
The first layer is the input layer, the last, the output layer
Layers between the input and output layers are called hidden
Excitatory influence
Examples of feed-forward
topologies
4-node input layer
sum w1 x1 ... wn xn
Here xi is the signal and wi is the weight on
connection I and n is the number of input signals
Linear
1
1 e ( sum c )
Activation function
graphs
Linear
0
1
Step
0
1
Sigmoid
0
Example
Step activation function with c = 0.38
0
w1 = 0.3
y =1
w2 = 0.5
y =1
y =1
w3 = - 0.1
y =1
Classification
Pattern association
Learning laws
Representation of training
examples
Description attributes
local encoding
Class attribute
Network configuration
Pyramid structure
the first hidden layer fewer nodes than the input layer
each hidden layer has less than its predecessor
the output layer has least
Training epochs
Conversion of attributes:
Attribute
Outlook
Temperature
Humidity
Windy
Values
sunny, overcast, rain
-50 to 150 C
low, normal, high,
true, false
Class
yes, no
Attribute
Sunny
Overcast
Rain
Temperature
Low
Normal
High
Windy
Values
0, 1
0, 1
0, 1
0 to 1
0, 1
0, 1
0, 1
0, 1
Play golf
1, 0
(T+50)/200
Network configuration
Use a single layer network (no hidden units) with step function
to illustrate the delta rule
Set = 0.1
Sunny
Overcast
Rain
Temperature
Low
Normal
High
Windy
w1
w2
w3
w4
w5
w6
w7
w8
-1
w0=0.3
(bias)
w1 =
w2 =
w3 =
w4 =
w5 =
w6 =
w7 =
w =
-0.5
-0.4
0.2
0.3
0.1
-0.1
-0.2
0.4
Feeding a training
example
encoded as:
(1, 0, 0, 0.355, 1, 0, 0, 0) : 1
T = 1 thus error is T - O = 1 - 0 = 1
Using delta rule change weights by:
wi = (T-O) xi
= 0.1 * 1 * xi
= 0.1 * xi
Learning in multi-layered
networks
The Backpropagation
process
error 1
error 2
error 3
wk
error k
error_contribution = w1 * error 1 + + wk * error k
The backpropagation
algorithm
A backpropagation network is
Backpropagation algorithm
1. Initialise all network weights to small random numbers
(between -0.05 and 0.05)
2. begin epoch
for each training example do:
propagate input to output layer;
from output layer, backpropagate errors;
update weights
end epoch
3. If termination condition is met, stop else goto 2
Termination conditions
Backpropagation as a
search
Backpropagation versus
ID3
These two algorithms are the giants of
classification learning
Which is better?