Professional Documents
Culture Documents
Design of Interchannel MRF Model For Probabilistic Multichannel Image Processing
Design of Interchannel MRF Model For Probabilistic Multichannel Image Processing
where the Bayes’ rule is used. Since is the observation of will show a sharper peak than other terms. Therefore, the max-
, and this can be further reduced to imum of (10) will occur when is close to the optimal solution
of the single-channel method, that is,
(6)
(11)
from the Markovian property. The first part of the cost function
comes from observation models that depend on applications [3], Moreover, by assuming this approximation is accurate, we can
[8]. Hence, we need to design for a new MAP-MRF estimate the th channel image as
estimator.
(12)
Although we focus on the development of conditional prob- Since (11) is a conventional (single-channel) problem and (13)
ability by assuming that a noise-free reference is available, is the same as (6), our method based on (6) can address the joint
it should be noted that the joint probability is a estimation problem, although it may not be optimal. Note that
more powerful model. From the joint probability, we can get this approximation holds only when the th channel is much
the marginal probability and the conditional probability informative, and we have to solve (13) after (11).
. Also, it enables joint estimation. In order words, we Fortunately, there are many algorithms showing improved
can estimate and simultaneously from their degraded performance with less complexity compared to MRF-based
observations ( and ) by maximizing the posterior prob- methods, and we can use them for the first subproblem (11).
ability of them as To be specific, wavelet-based denoising methods such as [23]
show better denoising performance than MRF-based methods
[8]. Also, it is known that the edge-directed image interpolation
(7) methods outperform MRF-based methods [12], [24]. In the
(8) experimental section, we have used this greedy method in order
to handle nonideal references, and the results show that cor-
(9) rupted references can also improve the performance of another
channel processing as long as it is informative. In summary, (6)
Here, the Bayes’ rule and the Markovian property are used, and is an effective formulation that can cover many interesting ap-
the subscripts “ ” indicate that they are the results of the joint plications, and the design of conditional probability
estimation (we use this notation in order to differentiate them allows the estimation.
from the results of other estimators). However, the design of Finally, it is worth noting that we can construct the joint prob-
is believed to be a very difficult task, since we have ability by multiplying our conditional probability with the con-
to consider the spatial correlations and the channel correlations ventional image priors [3], [5], [8]
simultaneously. It is well known that even the learning of spatial
(14)
correlations (i.e., the learning of the marginal probability) is a
difficult and very time-consuming task [8], [10]. Moreover, its
However, we have found that this model is not a practical one
extension to multichannel images having more than two chan-
due to its computational complexity, which will be discussed in
nels (i.e., ) is not straightforward.
Section IV.
Hence, rather than developing the joint probability and opti-
mizing equation (9), we try to develop a relatively simple model
and focus on the optimization of (6). Fortunately, this approach III. CHANNEL MODEL
is effective in many interesting cases. Almost noise-free refer- Here, we review conventional channel models and propose
ence images are available in satellite image fusion application a new probabilistic model in order to represent the probability
[20] and night-vision applications [21], so that (6) is a suffi- of given its reference . We assume that the resolution of
cient formulation. Even though noise-free reference images are each channel image is the same (their observations may have
not available, we can develop a suboptimal method that can different resolutions) and they are defined on , which is a set of
handle nonideal references. For the explanation of our subop- sites (pixels) as shown in Fig. 2. Also, we denote the positional
timal method, let us rewrite (9) as vector at the site as , and indicates the
pixel value of the th channel image at .
(18)
(16)
(20)
Since the luminance-chrominance representation (e.g., YCbCr,
where is a vector representing the parameters of
YUV) is simply an affine transformation of RGB representation,
and , and is the number of sites (pixels) in , i.e.,
the relationship of and in such color spaces is given
. From the MRF-GRF equivalence, the probability of
by
, given and , can be expressed as
(17)
(21)
where and are two functions defined on the image do-
main . Since the estimation of and is a severely where is a partition function. If we consider as a random
ill-posed problem, several assumptions on two functions were field, we can get
posed in previous work. In chrominance interpolation and de-
mosaicking, two functions are modeled as piecewise constant (22)
functions by assuming that albedo is a constant in the same ob-
ject (material). Under the assumption, we can estimate and
from some true pixel values, and missing pixel values can which is intractable due to infinitely many instances of . In
be computed from them. However, such an approach is appli- order to make it tractable, we assume that the prior is non-in-
cable only to limited applications, because the parameter esti- formative (i.e., and shows
mation is sensitive to noise. a very sharp peak at it optimal estimate , so that
. Here, the estimate can be found
B. Proposed Probabilistic Channel Model by solving least squares problems. In summary, the conditional
probability is given by
In many cases, and are slowly varying within an
object. Even if they may experience abrupt changes at object (23)
boundaries, these changes are not modeled as step functions but
KOO AND CHO: DESIGN OF INTERCHANNEL MRF MODEL FOR PROBABILISTIC MULTICHANNEL IMAGE PROCESSING 605
(28)
Fig. 3. Three image patches from different channels. (a) I . (b) I . (c) I .
Since corresponding pixels in I and I show a linear pattern, 8(c ) is small, (29)
indicating that I is a probable one when I is given. On the other hand, 8(c )
corresponding I and I is large and we can expect that I is an unlikely patch
when I is its reference. (30)
(26)
(27)
(34)
TABLE I
SUMMARY OF THE PROPOSED METHOD
(36)
2
Fig. 7. Illustration of the proposed fusion process. The resolution of I is 4
2
higher than that of K . For example, the size of K is 128 128 and the size
2
of I is 512 512.
(41) Fig. 8. One of image sets used in the image fusion experiments. (a) PAN image
2 2 2
(512 512). (b) R image (128 128). (c) G image (128 128). (d) B image
2
(128 128).
where is the variance of observation noise, is a raster-scan
ordered pixel vector of the low-resolution image is from
high-resolution image by sequentially placing each of 4 4 TABLE II
PSNR COMPARISON OF IMAGE FUSION
blocks, and is a filtering and decimation matrix:
..
.
(42)
where is scaling factor (in this case, , see [3] for more
details). Note that we can use this model because images are
taken at the same viewpoint with the same shot angle as shown Fig. 8) and summarize the result in Table II. As can be seen,
in Fig. 8. Otherwise, multisensor image registration methods our method yields more than 3-dB higher PSNR compared with
should be applied [31]. the IHS method and 1-dB higher PSNR compared with the
We can get the result by minimizing the cost function using AWT method. Moreover, the proposed method suppresses color
the conjugate gradient algorithm [27] distortion which is the major drawback of previous methods.
Figs. 9 and 10 show that the proposed method introduces
relatively small amount of color distortion, and does not suffer
(43)
from ringing artifacts which are observed in AWT method [e.g.,
Fig. 9(c)].
where the first term encodes the condition that the interpolated
result should be consistent with the observation. On the other C. Chrominance Image Interpolation With Denoising
hand, the second term in (43) prevents the structure misalign- Chrominance image processing has received relatively less
ment between the interpolated one and the reference . Since attention and was independently treated in most applications
becomes small only when has sim- [22]. However, by considering the luminance images having
ilar structure to (as illustrated in Fig. 3), the minimization of higher SNR and/or higher resolution in the processing of
(43) naturally results in which is consistent with the refer- chrominance signals, the performance is expected to be im-
ence image. Note that, in our method, interpolation methods are proved. A typical example may be the conventional image/video
not explicitly used. Rather, the overall process (an interpolation standards with 4:2:0 format: the chrominance channel images
process considering a reference image as well as an observed (e.g., Cb and Cr channels) have low resolution and low SNR
image) is modeled as a single optimization problem. compared with luminance channel images. Hence, we can
For the objective evaluation of our method, we measure apply the proposed framework to chrominance image denoising
the PSNR of enlarged images (one of image sets is shown in and interpolation (note that our method cannot be applied to the
KOO AND CHO: DESIGN OF INTERCHANNEL MRF MODEL FOR PROBABILISTIC MULTICHANNEL IMAGE PROCESSING 609
Fig. 10. Subjective comparison of satellite image fusion. (a) Original image.
(b) IHS method. (c) AWT method. (d) Proposed method.
standards that treat RGB components exactly in the same way, Fig. 12. Denoising and interpolation experiments on chrominance channel:
(a) Original Cb channel of Pepper. (b) Corrupted Cb (PSNR = 25:0 dB).
e.g., RGB 4:4:4 format [32]). In this application, means a (c) BLS-GSM + bicubic interpolation PSNR = 35.03 dB. (d) Proposed
chrominance channel image and means a luminance channel method (PSNR = 35.97 dB).
image. The cost function is similar to that of the satellite
image fusion application because this application also uses the
observation model in [3] . luminance image by using BLS-GSM algorithm [23], and then,
Since a noise-free reference channel is not available (only our method is applied. The proposed method is compared with
its noisy observation is available in practice), we also adopt the the conventional channel independent bicubic interpolation for
greedy method presented in Section II-C: we first denoise the Lena and Pepper images as shown in Figs. 11 and 12. For fair
610 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 20, NO. 3, MARCH 2011
[29] B. Aiazzi, L. Alparone, S. Baronti, and A. Garzelli, “Context-driven Nam Ik Cho (S’86–M’92) received the B.S., M.S.,
fusion of high spatial and spectral resolution images based on oversam- and Ph.D. degrees in control and instrumentation
pled multi-resolution analysis,” IEEE Trans. Geosci. Remote Sens., vol. engineering from Seoul National University, Seoul,
40, no. 10, pp. 2300–2312, Oct. 2002. Korea, in 1986, 1988, and 1992, respectively.
[30] T. Ranchin, B. Aiazzi, L. Alparone, S. Baronti, and L. Wald, “Image fu- From 1991 to 1993, he was a Research Associate
sion-the arsis concept and some successful implementation schemes,” with the Engineering Research Center for Advanced
ISPRS J. Photogramm. Remote Sens., vol. 58, pp. 4–18, 2003. Control and Instrumentation, Seoul National Univer-
[31] M. Irani and P. Anandan, “Robust multi-sensor image alignment,” in sity, Seoul, Korea. From 1994 to 1998, he was with
Proc. Int. Conf. Comput. Vis., Jan. 1998, pp. 959–966. the University of Seoul, Seoul, Korea, as an Assis-
[32] B. C. Song, Y. G. Lee, and N. H. Kim, “Block adaptive inter-color tant Professor of Electrical Engineering. He joined
compensation algorithm for RGB 4:4:4 video coding,” IEEE Trans. the School of Electrical Engineering, Seoul National
Circuit Syst. Video Technol., vol. 18, no. 10, pp. 1447–1451, Oct. 2008. University, in 1999, where he is currently a Professor. His research interests in-
clude image processing and adaptive filtering.
Hyung Il Koo (S’09–M’10) received the B.S., M.S.,
and Ph.D. degrees in electrical engineering and com-
puter science from Seoul National University, Seoul,
Korea, in 2002, 2004, and 2010, respectively.
Currently, he is a Senior Research Engineer with
Qualcomm R&D Center, Seoul, Korea, where he
is responsible for developing core technology in
computer vision. His research interests include
image processing, pattern recognition, and computer
vision.