Professional Documents
Culture Documents
Artificial Neural Network
Artificial Neural Network
NETWORKS
Biological Neuron to Artificial Neuron
McCulloh Pitts Perceptron Model
Layer of Neurons
Activation Function
Artificial Learning
Types of Learning
Introduction to Back Propagation Networks
Applications of Neural Network
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
The human nervous system can be broken down into three stages that
may be represented in block diagram form as:
ANNs have been developed as generalizations of mathematical models of neural biology, based on the
assumptions that:
Wa
X1
Wb Y
X2 f()
Wc
X3
• It is important to note that a neuron can send only one signal at a time, although that signal
is broadcast to several other neurons.
• Neural networks are configured for a specific application, such as pattern recognition or
data classification, through a learning process
• In a biological system, learning involves adjustments to the synaptic connections between
neurons
same for artificial neural networks (ANNs)
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Artificial Neural Network
Synapse Nucleus
x1 w1
y
Axon
x2 w2 Activation Function:
yin = x1w1 + x2w2 (y-in) = 1 if y-in >=
and (y-in) = 0
Dendrite
- A neuron receives input, determines the strength or the weight of the input, calculates the total weighted input,
and compares the total weighted with a value (threshold)
- The value is in the range of 0 and 1
- If the total weighted input greater than or equal the threshold value, the neuron will produce the output, and if
the total weighted input less than the threshold value, no output will be produced
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Dendrites: Input
Cell body: Processor
Synaptic: Link
Axon: Output
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Analogy of ANN with BNN
• The dendrites in biological neural network is analogous to the weighted inputs based on
their synaptic interconnection in artificial neural network.
• Cell body is analogous to the artificial neuron unit in artificial neural network which also
comprises of summation and threshold unit.
• Axon carry output that is analogous to the output unit in case of artificial neural network.
So, ANN are modelled using the working of basic biological neurons.
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Comparing ANN with BNN
As this concept borrowed from ANN there are lot of similarities though
there are differences too.
• Similarities are in the following table
Biological Neuron Artificial Neuron
Cell Neuron
Axon Output
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Criteria BNN ANN
The cycle-time of execution in few The cycle-time of execution in few
Speed
milliseconds nanoseconds
Massively parallel, slow but superior than Massively parallel, fast but inferior than
Processing
ANN BNN
1011 neurons and 1015 interconnections. The The size and complexity of an ANN is
Size &
size and complexity of a BNN is more than based on the chosen application and the
Complexity
an ANN network designer.
Very precise, structured and
Learning They can tolerate ambiguity formatted data is required to
tolerate ambiguity
Performance degrades with even partial It is capable of robust performance, hence
Fault tolerance
damage has the potential to be fault tolerant
Stores the information in
Storage Stores information in its interconnection or
continuous memory locations, loss of
capacity in synapse. No Loss of memory
memory may happen sometimes.
Control There is no such control unit for monitoring
It is very simple as compared to BNN.
mechanism in the brain
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
ANN possesses the following characteristics:
• It is a neurally implemented mathematical model
• There exist a large number of highly interconnected processing elements called
neurons in an ANN.
• The interconnections with their weighted linkages hold the informative
knowledge.
• The input signals arrive at the processing elements through connections and
connecting weights.
• The processing elements of the ANN have the ability to learn, recall and
generalize from the given data by suitable assignment or adjustment of weights.
• The computational power can be demonstrated only by the collective behavior of
neurons, and it should be noted that no single neuron carries specific information.
These make the ANNs as connectionist models, parallel distributed processing
models, self-organizing systems, neuro-computing systems and neuro-morphic
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Evolution of Neural Networks
1943 McCulloch-Pitts neurons
1949 Hebb’s law
1958 Perceptron (Rosenblatt)
1960 Adaline, better learning rule (Widrow, Huff)
1969 Limitations (Minsky, Papert)
1972 Kohonen nets, associative memory
1977 Brain State in a Box (Anderson)
1982 Hopfield net, constraint satisfaction
1985 ART (Carpenter, Grossfield)
1986 Backpropagation (Rumelhart, Hinton, McClelland)
1988 Neocognitron, character recognition (Fukushima)
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
Important terminologies of ANNs
• Weights
• Bias
• Threshold
• Learning rate
• Momentum factor
• Vigilance parameter
• Notations used in ANN
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
• Weights:
Each neuron is connected to every other neuron by means of directed links
Links are associated with weights
Weights contain information about the input signal and is represented as a matrix
Weight matrix also called connection matrix
• Bias:
Bias is like another weight. Its included by adding a component x0=1 to the input vector X.
X=(1,X1,X2…Xi,…Xn)
Bias is of two types: Positive bias: increase the net input, Negative bias: decrease the net input
• The relationship between input and output given by the equation of straight line y=mx+C
C(bias)
Input X Y y=mx+C
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
• Threshold:
Set value based upon which the final output of the network may
be calculated
Used in activation function
The activation function using threshold can be defined as
1ifnet
• Learning rate: f ( net )
Denoted by α. 1ifnet
Used to control the amount of weight adjustment at each step of
training
Learning rate ranging from 0 to 1 determines the rate of
learning in each time step
BIOLOGICAL NEURON TO ARTIFICIAL NEURON
• Momentum factor:
– used for convergence when momentum factor is added to
weight updation process.
– Used in back propagation network.
• Vigilance parameter:
– Denoted by ρ
– Used to control the degree of similarity required for patterns
to be assigned to the same cluster
– It range approximately from 0.7 to 1 to perform useful work
in controlling the no. of clusters.
MC-CULLOH PITTS PERCEPTRON MODEL
The early model of an artificial neuron is introduced by Warren McCulloch and
Walter Pitts in 1943.
It is usually called as M-P neuron. The M-P neurons are connected by directed
weighted paths. The activation of a M-P neuron is binary, that is, at any time
step the neuron may fire or may not fire. The weights associated with the
communication links may be excitatory (weight is positive) or inhibitory
(weight is negative). All the excitatory connected weights entering into a
particular neuron will have same weights.
The threshold plays a major role in M-P neuron: There is a fixed threshold for
each neuron, and if the net input to the neuron is greater than the threshold then
the neuron fires. Any nonzero inhibitory input would prevent the neuron from
firing. The M-P neurons are most widely used in the case of logic function.
MC-CULLOH PITTS PERCEPTRON MODEL
McCulloch-Pitts neuron model.
A simple M-P neuron is shown in Figure. As already discussed, the M-P neuron has both
excitatory and inhibitory connections. It is excitatory with weight (w > 0) or inhibitory with
weight –p (p < 0). In Figure inputs from X1 to Xn possess excitatory weighted connections
and inputs from Xn+ 1 m Xn+m possess inhibitory weighted interconnections. Since the firing of
the output neuron is based upon the threshold, the activation function here is defined as
For inhibition to be absolute, the threshold with the activation function should satisfy the
MC-CULLOH PITTS PERCEPTRON MODEL
The output will fire if it receives say “k” excitatory inputs but no inhibitory
inputs, where
The M-P neuron has no particular training algorithm. An analysis has to be
performed to determine the values of the weights and the threshold. Here the
weights of the neuron are set along with the threshold to make the neuron
"perform a simple logic function. The M-P neurons are used as building blocks
on which we can model any function or phenomenon, which can be represented
as a logic function
Features of McCulloch-Pitts model
• Allows binary 0,1 states only
• Operates under a discrete-time assumption
• Weights and the neurons’ thresholds are fixed in the model and no
interaction among network neurons
• Just a primitive model