Professional Documents
Culture Documents
Farkas Image Classif NN
Farkas Image Classif NN
Igor Farkaš
2012
Deep architectures
● in biological systems
nonlinear transformations →
deep learning
Methods using artificial neural networks
● brain-inspired
● basic building blocks (computing elements) – artificial neurons:
● deterministic (perceptron, RBF) → discriminatory models (c)
● stochastic (probabilistic) → generative models (a,b)
supervised (with teacher) unsupervised (selforganized)
reinforcement learning (partial feedback)
Typical artificial neuron model
Deterministic model
y = f (∑i wixi )
Stochastic model P(s=+1) = 1/(1+exp(∑i wixi + ))
Discrete perceptron
(Rosenblatt, 1962)
y = f(∑nj=1 wj xj )
n1
y= f ∑ j=1 w j x j x n1 =−1
Training:
1. choose input x, compute output y, set E = 0
2. evaluate error function e(t) = ½ (d – y)2, E ← E + e(t)
3. adjust weights using delta rule (if e(t) > 0)
4. if all patterns used, then goto 5, else go to 1
5. if E = 0 (all patterns in the set classified correctly), then end
else reorder inputs, E ¬ 0, go to 1
Perceptron classification capacity
x2
2D
example
linear
decision
boundary
x1
●
Inputs x , weights w, v, outputs y
●
Nonlinear activation function f
● Unit activation:
n1
h k = f ∑ j=1 v kj x j
q1
yi = f ∑ k=1 w ik h k
● Bias input: x n1=h q 1=−1
● Activation function examples:
f(net) = 1 / (1+exp(net))
f(net) = tanh(net)
Learning equations for original BP
Hidden-output weights:
wik(t+1) = wik(t) + i hk where i = (di – yi) fi'
Input-hidden weights:
vkj(t+1) = vkj(t) + k xj where k = (Σi wik i) fk'
(Smolensky, 1986)
(Hinton, 2007)
(Hinton, 2007)
(Hinton, 2007)
Deep Belief Network (DBN) = stacked RBMs
(Hinton, 2006)
Convolutional DBN
● sharing weights
among all locations
within a layer
● Probabilistic max-pooling
● Both features lead to
translational invariance
and contribute to scalability
Trained by error
back-propagation!
“where”
“what”