Professional Documents
Culture Documents
Image Fusion Using Wavelet Transform: Zhu Shu-Long
Image Fusion Using Wavelet Transform: Zhu Shu-Long
SIPT UCI ACSG Table des matières Index des auteurs Recherches Sortir
Zhu Shu-long a, b
a
Department of Remote Sensing Information Engineering, Zhengzhou Institute of Surveying & Mapping, Zhengzhou, China,
Postcode: 450052 -zsl1964@public2.zz.ha.cn
b
LIESMARS of Wuhan University, Wuhan, China, Postcode: 430079
KEY WORDS: Multi-source images, Fusion, Wavelet transform, Edge processing, Wavelet based fusion, Fused Images
ABSTRACT:
Extracting more information from multi-source images is an attractive thing in remotely sensed image processing, which is recently
called the image fusion. In the paper, the image fusion algorithm based on wavelet transform is proposed to improve the geometric
resolution of the images, in which two images to be processed are firstly decomposed into sub-images with different frequency, and
then the information fusion is performed using these images under the certain criterion, and finally these sub-images are
reconstructed into the result image with plentiful information.
and imagery, and been proved to be an effective tool to process is the orthogonal complement of V j with respect to V j + 1 ,
the signals in multi-scale spaces ( Thierry Ranchin et al. ). In i.e. V j +1 =V j + W j
, ψ ( x ) is wavelet function of
this paper , the image fusion algorithm based on wavelet
subspace W 0 . If a signal f ( x ) ∈ V j +1
,then it can be
transform is proposed to improve the geometric resolution of
the images, in which two images to be processed are firstly expressed as :
f (x ) = ∑c n
n
j
ϕ j, n + ∑d n
j
ψ j ,n
(2) reconstruction formula of signal. The formulas (4)、(5) and (6)
are :
∑c
n
n
j +1
ϕ j +1, n
=
∑cn
n
j
ϕ j, n + ∑d n
j
ψ j ,n
(3)
1
c mj , n =
2
∑
k ,l ∈ Z
c kj ,+l 1 h k − 2 m h l − 2 n
1
c kj = ∑c n
j +1
ϕ j +1,n ,ϕ j ,k =
1
∑h n−2k c nj + 1 (4) d mj 3,n =
2
∑
k , l∈ Z
c kj ,+l 1 g k − 2 m g l − 2 n
n 2 n
Using the same method, we also have: and its reconstruction formula is :
1
d kj =
2
∑ g n − 2 k c nj + 1 (5)
n
1 ~ ~ ~ ~
cmj+,n1 = ∑ckj,l h2k−mh2l−n + ∑dkj,1l h2k−mg~2l−n + ∑dkj,2l g~2k−mh2l−n + ∑dkj,3l g~2k−mg~2l−n
2k,l∈Z k,l∈Z k,l∈Z K,L∈z
The formulas (4) and (5) are the decomposition formula of
signal, where c j
is an approximation of c j +1
, and d j
is (8)
j +1
the detailed part of c .
where c j
is an approximation of c j +1
, d j1
、d j2
and
If having the decomposed signals {c n
j
, n ∈ Z } and j3
are the detailed parts of c j +1
. Fig. 1 illustrates the
d
{d n
j
, n ∈ Z },then two sides of formula (3) are multiplied Mallat algorithm of 2D images intuitively.
by ϕ j +1,k
, and then integrated with respect to x , we can
obtained :
j +1
forward transform cj d j1
c
backward transform
c nj + 1 = ∑ c nj ϕ d j2 d j3
k
j ,k ,ϕ j +1, n + ∑ k
d kj ψ j ,k ,ϕ j +1,n
1 ~
= ∑ c kj h 2 k − n + ∑d j
g~ 2 k − n
2 k
k
2.2 Edge Processing for Wavelet Transform
(6)
Since our image fusion algorithm need to use the forward and forward wavelet transform, which have the same resolution at
the inverse algorithm of wavelet transform, in order to the same levels and fifferent resolution among different levels;
reconstruct the signals precisely, the image edges must be and (3) information fusion is performed based on the
processed under the process of wavelet transform, otherwise high-frequency subimages of decomposed images; and finally
some information will be lost in this operation. the result image is obtained using inverse wavelet transform.
Assuming that the low-pass filter and the corresponding Let A( x, y ) and B ( x, y ) be the images to be fused,the
high-pass filter are four coefficient Daubechies' ( I. Daubechies, decomposed low-frequency sub-images of A( x, y ) and
1988), i.e. H = (h0 , h1 , h2 , h3 ) and
B ( x, y ) be respectively lAJ ( x, y ) and lB J ( x, y )( J
,the decomposed high-frequency
is the parameter of resolution)
G = ( g − 2 , g −1 , g 0 , g 1 ) respectively, the samples of the
sub-images of A( x, y ) and B ( x, y ) be
original discrete signal are C 0 = (c00 , c10 ,...,c M0 −1 ) , then
respectively hA ( x, y ) and hB j ( x, y ) ( j is the
k k
j
the image edges are processed according the following precise
parameter of resolution and j = 1,2, L , J . for every j ,
formulas ( Zhu Shu-long and Geng zhe-Xun, 1999 ) in our k = 1,2,3 ) , the gradient image generated from
experiments: hA kj ( x, y ) and hB kj ( x, y ) be respectively
GAkj ( x, y ) and GB kj ( x, y ) , then the fused
For left image edges, we have:
high-frequency sub-images F j ( x , y ) are:
k
if GA j ( x, y ) > GB j ( x , y )
k k
c −01 = c −02 = 0
F jk ( x, y ) = hA kj ( x, y ) ,
1 h2 1
if GA j ( x, y ) < GB j ( x , y )
k k
c −1 = h d 0
F jk ( x, y ) = hB kj ( x, y ) ,
1
FJ ( x, y ) = k1 ⋅ lAJ ( x, y ) + k 2 ⋅lB J ( x, y )
For right image edges, we have:
where k1 and k 2 are given parameters, if image B is
The main idea of our algorithm is that: (1) the two images to be
For multi-spectral images, we can use the same method
processed are resampled to the one with the same size; and (2)
proposed above to process every band images respectively. For
they are respectively decomposed into the subimages using
color images, we can convert the color images from RGB ( i.e.
Red, Green, Blue ) space to HIS ( i.e. Hue, Intensity,
Saturation ) space, and then only process the intensity 5. THE EXPERIMENTS
component of HIS space using the same method proposed
above, and finally acquired the fused the images. In our experiments, two images are used to verify our algorithm,
the first is the image composed of SPOT multi-spectral images
with the resolution of 20 m ( as shown in Fig.2 ), the second is
4. THE ASSESSMENT FOR THE FUSED IMAGES SPOT panchromatic image with the resolution 10 m ( as shown
in Fig.3 ). In order to fully utilize the spectral information of
The index "Average gradient" is applied to measure the detailed the former and geometric information of the latter, the image
information in the images. The formula of average gradient "g" fusion algorithm proposed above is used to introduce the details
can be denoted as follows: of the latter into the former so that the result image ( as shown
in Fig.4 ) has not only good spectral information but also good
geometric information. The table 1 illustrates the average
∂f (x, y) ∂f (x, y)
2 2
M−1N−1
1
g= ∑∑ +
(M −1) ⋅ (N −1) x=1 y=1 ∂x ∂y
(11) gradient "g" in these three images quantitatively, in which the
average gradient of the result image is obviously greater than
that of the original image 1.
Fig. 2. The original image 1 with Fig. 3. The original image 2 with Fig. 4. The result image generated by
the resolution 20 m the resolution 10 m our algorithm
Zhu Shu-long and Geng zhe-Xun, 1999, The Wavelets and Its
Applications to Graphics and Imagery, The PLA Press, P.R.C.