Professional Documents
Culture Documents
An Introduction To Independent Component Analysis (ICA)
An Introduction To Independent Component Analysis (ICA)
(ICA)
吳育德
陽明大學放射醫學科學研究所
台北榮總整合性腦功能實驗室
The Principle of ICA: a cocktail-party problem
Since mean and variance of xk can grow without bound as k, consider
instead of xk the standardized variables
xk m x k
yk
xk
The distribution of yk a Gaussian distribution with zero mean and unit varianc
e when k.
How to estimate ICA model
• Principle for estimating the model of ICA
Maximization of NonGaussianity
Measures for NonGaussianity
• Kurtosis
Kurtosis : E{(x- )4}-3*[E{(x-)2}] 2
1
T
z Vx D E x 2
T T
E{zz } VE1{xx }V T
1
D 2 ET EDET ED 2
I
Importance of whitening
For the whitened data z, find a vector w such that the linear c
ombination y=wTz has maximum nongaussianity under the con
strain E{ y 2 } 1
Then
1 E{ y 2 }
E{w T z z T w}
w E{z z }w
T T
w w T
F (w )
[ 2w ] 0
w
F (w )
2 w
w
At the stable point, the gradient of F(w) must point in the direct
ion of w, i.e. equal to w multiplied by a scalar.
Gradient of kurtosis
F (w ) E{(w z ) } 3E{(w z ) }}
T 4 T 2
w w
1 t 1 (w T z (t )) 4 3(w T w ) 2
T
1 T
T E{y} y(t )
T t 1
w
4 T
T
t 1
z (t )[ w T
z (t )]3
3 * 2( w T
w)(w w )
2
4sign(kurt(w z))[ E{z[w z] } 3w w ]
T T 3
Fixed-point algorithm using kurtosis
F ( w )
wk+1 = wk + k
w
k
-1 F ( w )
=( + ) k
2 w k
2
Therefore, w E{z[w T z ] } 3w w
3
w w/ w
Note that adding the gradient to wk does not change its direction, since
wk+1 = wk - ( 2 wk )
= (1- 2 ) wk
1. Centering x~
x m ~x
T 2
5. Let w p E{z[ w p z ] } 3w p w p
3
6. Do deflation decorrelation p 1
w p w p (w Tp w j )w j
j 1
7. Let wp wp/||wp||
Approximation of negentropy
Fixed-point algorithm using negentropy
Entropy
H (y ) p y ( ) log p y ( )d 0
Negentropy
J (y ) H (y gauss ) H (y ) 0
Approximation of negentropy
J ( y ) [ E{G ( y )} E{G ( )}]2
1 g1 ( y ) tanh a1 y
G1 ( y ) log cosh a1 y, 1 a1 2
a1
G2 ( y ) exp( y 2 / 2) g 2 ( y ) y exp( y 2 / 2)
y4
G3 ( y ) g3 ( y) y 3
4
Fixed-point algorithm using negentropy
Max J(y)
w E{zg(wTz)} E{g '(wTz)} w
w w/||w||
Convergence : |<wk+1 , wk>|=1
1
G1 ( y ) log cosh a1 y, 1 a1 2 g1 ( y ) tanh a1 y g1 ( y ) a1[1 tanh 2 (a1 y )]
a1
G2 ( y ) exp( y 2 / 2) g 2 ( y ) y exp( y 2 / 2) g 2 ( y ) (1 y 2 ) exp( y 2 / 2)
y4
G3 ( y ) g3 ( y) y 3 g 3 ( y ) 3 y 2
4
Fixed-point algorithm using negentropy
1. Centering x~
x m ~x
5. Let w p E{z g ( w T
p z )} E{ g ( w T
p z )}w p
6. Do deflation decorrelation p 1
w p w p (w Tp w j )w j
j 1
7. Let wp wp/||wp||
8. If wp has not converged, go back to step 5.
9. Set p p+1. If p m, go back to step 4.
Implantations
• Create two uniform sources
Implantations
• Create two uniform sources
Implantations
• Two mixed observed signals
Implantations
• Two mixed observed signals
Implantations
• Centering
Implantations
• Centering
Implantations
• Whitening
Implantations
• Whitening
Implantations
• Fixed-point iteration using kurtosis
Implantations
• Fixed-point iteration using kurtosis
Implantations
• Fixed-point iteration using kurtosis
Implantations
• Fixed-point iteration using negentropy
Implantations
• Fixed-point iteration using negentropy
Implantations
• Fixed-point iteration using negentropy
Implantations
• Fixed-point iteration using negentropy
Implantations
• Fixed-point iteration using negentropy
Fixed-point algorithm using negentropy
Entropy
H (y ) p y ( ) log p y ( )d
Negentropy
J (y ) H (y gauss ) H (y )
Approximation of negentropy
Fixed-point algorithm using negentropy
High-order cumulant approximation
1 1
J ( y) E{ y 3 } kurt ( y ) 2
12 48
Replace the polynomial function by any n
onpolynomial fun Gi
ex.G1 is odd and G2 is even
J ( y ) k1 ( E{G1 ( y )}) 2 k 2 ( E{G2 ( y )} E{G2 ( )}) 2
It’s quite common that most r.v. have approximately symmetric dist. E{G1 ( y )} 0
w [ E{zg (w T z )}]
w w/ w
Fixed-point algorithm using negentropy
w [ E{zg (w T z )}]
w w/ w
The iteration doesn't have the good convergence properties because the
nonpolynomial moments don't have the same nice algebraic properties.
(1 )w [ E{zg (w T z )}] w
w w/ w
Finding by approximative Newton method
1
G1 ( y ) log cosh a1 y, 1 a1 2 g1 ( y ) tanh a1 y g1 ( y ) a1[1 tanh 2 (a1 y )]
a1
G2 ( y ) exp( y 2 / 2) g 2 ( y ) y exp( y 2 / 2) g 2 ( y ) (1 y 2 ) exp( y 2 / 2)
y4
G3 ( y ) g3 ( y) y 3 g 3 ( y ) 3 y 2
4