Professional Documents
Culture Documents
Neuro Fuzzy - Session 2
Neuro Fuzzy - Session 2
SESSION 2
An Introduction
• There are certain tasks which a computer can do better than humans
• Multiplying 2 big numbers
• Search through millions of records
An Introduction
• There are certain tasks where humans outperform
computers
• We don’t see many robots driving the car on the road
• Some companies are trying driverless cars
• But it is still in initial phase
• Natural language conversation is also an area where
humans outperform machines
• Plumbing, Electrical Works and various other expert
works
An Introduction
• Machine learning tries to make computers
better at doing things that humans
traditionally can do better than machines
• Make machines learn things like humans do
• We need to understand how human brain
works
• Whenever we think or make a decision a signal
is generated and the neurons light up
An Introduction
𝑚 𝑤11 𝑤1𝑛
⋯
𝑤21 𝑤2𝑛
𝑦_𝑖𝑛𝑛 = 𝑥1 𝑤1𝑛 + 𝑥2 𝑤2𝑛 + … + 𝑥𝑚 𝑤𝑚𝑛 = 𝑥𝑖 𝑤𝑖2 ⋮ ⋱ ⋮
𝑖=1 𝑤𝑚1 ⋯ 𝑤𝑚𝑛
𝑌_𝑖𝑛 = 𝑋 × 𝑊
2. MultiLayer Feed Forward ANN
• Similar to single layer feed forward net except that there is one or more additional
layers of processing units between the input and the output layers
• The additional layers are called the hidden layers of the network
MultiLayer Feed Forward ANN – cont..
• Net input to the hidden layer
𝑌_𝑖𝑛 = X × 𝑊
• Net input to the output layer
Z𝑖𝑛 = Y_out × 𝑉
Where
𝑋 = [𝑥1, 𝑥2, … , 𝑥𝑚] is the i/p vector
Y_in= [𝑦_𝑖𝑛1, 𝑦_𝑖𝑛2, … , 𝑦_𝑖𝑛𝑛] is the
net i/p vector to the hidden layer
Z_in= [𝑧_𝑖𝑛1, 𝑧_𝑖𝑛2, … , 𝑧_𝑖𝑛𝑟] is the
net i/p vector to the output layer
Y_𝑜𝑢𝑡 = [𝑦_𝑜𝑢𝑡1, 𝑦_𝑜𝑢𝑡2, … , 𝑦_𝑜𝑢𝑡𝑛]
is the o/p vector from the hidden layer
MultiLayer Feed Forward ANN – cont..
• It is defined as
1
𝑔 𝑥 =
1 + 𝑒 −σ𝑥
• The parameter σ is known as the steepness
parameter
• The transition from 0 to 1 could be made as
steep as desired by increasing the value of σ to • Derivative is similar to slope
appropriate extent • In a non-linear curve, at every point
• The first derivative of g(x), denoted by g’(x) is the slope is changing
Δ𝑦
• Derivative =
Δ𝑥
𝑔′ 𝑥 = σ𝑔 𝑥 (1 − 𝑔 𝑥 )
• It tells us how much the o/p
changes for the given change in i/p
Bipolar Sigmoid Function
′
σ
𝑔 𝑥 = 1+𝑔 𝑥 (1 − 𝑔 𝑥 )
2
Hyperbolic Tangent Function
• It is another widely employed bipolar activation
function
• It is closely related to the bipolar sigmoid function
• It is defined as
𝑒 𝑥 − 𝑒 −𝑥
ℎ 𝑥 = 𝑥
𝑒 + 𝑒 −𝑥
• Its first derivative is
ℎ′ 𝑥 = 1 + ℎ 𝑥 (1 − ℎ 𝑥 )
• When the input data is binary and not continuously
valued in the range from 0 to 1, they are generally
converted to bipolar form and then a bipolar sigmoid or
hyperbolic tangent activation function is applied on
them by the processing units