Download as pdf or txt
Download as pdf or txt
You are on page 1of 15

IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO.

2, FEBRUARY 2014 311

Adaptive Watermarking and Tree Structure


Based Image Quality Estimation
Sha Wang, Dong Zheng, Jiying Zhao, Member, IEEE, Wa James Tam, and Filippo Speranza

Abstract—Image quality evaluation is very important. In appli- image and the original image [1]–[7]. The widely used quality
cations involving signal transmission, the Reduced- or No-Refer- metrics in this category are the PSNR, wPSNR, Watson JND,
ence quality metrics are generally more practical than the Full- SSIM, etc. The Full-Reference quality metrics provide more
Reference metrics. In this study, we propose a quality estimation
method based on a novel semi-fragile and adaptive watermarking
accurate quality evaluation results comparing to the Reduced
scheme. The proposed scheme uses the embedded watermark to es- or No-Reference quality metrics. However, the Full-Reference
timate the degradation of cover image under different distortions. quality metrics become less practical when the original image is
The watermarking process is implemented in DWT domain of the not available. The Reduced-Reference quality metrics evaluate
cover image. The correlated DWT coefficients across the DWT sub- the quality of a distorted image using partial information of the
bands are categorized into Set Partitioning in Hierarchical Trees original image. In literature, such partial information can be
(SPIHT). Those SPHIT trees are further decomposed into a set of
bitplanes. The watermark is embedded into the selected bitplanes some features extracted from the original image [8]–[10]. The
of the selected DWT coefficients of the selected tree without causing Reduced-Reference quality metrics do not require the presence
significant fidelity loss to the cover image. The accuracy of the of the original image for quality evaluation. However, the
quality estimation is made to approach that of Full-Reference met- partial information of the original image need to be transmitted
rics by referring to an "Ideal Mapping Curve" computed a priori. to the receiver side either through an ancillary channel or by
The experimental results show that the proposed scheme can esti-
embedding into the transmitted image. The sacrifice of band-
mate image quality in terms of PSNR, wPSNR, JND and SSIM with
high accuracy under JPEG compression, JPEG2000 compression, width for transmitting the additional information needs to be
Gaussian low-pass filtering and Gaussian noise distortion. The re- considered. The No-Reference quality metrics estimate image
sults also show that the proposed scheme has good computational quality without accessing the original image [11]–[16]. In
efficiency for practical applications. practical applications, different quality metrics can measure the
Index Terms—Bitplane decomposition, complexity analysis, image degradation from different angle. For example, PSNR
DWT based watermark embedding, HVS masking, SPIHT tree measures image quality mathematically in terms of MSE; JND
structure, watermarking based image quality estimation. and SSIM intend to measure image quality with more emphasis
on the perceptual experience. Multiple metrics are often used
in practical applications to give a full coverage of the image
I. INTRODUCTION quality evaluation. In terms of applicability in signal transmis-
sion, Reduced- or No-Reference metrics are more preferable

T HE evaluation of image quality is very important in


today’s video broadcasting, transmission control, and
e-commerce, because quality is a key determinant of customer
over the Full-Reference metrics. To this end, a Reduced- or
No-Reference quality evaluation scheme which can evaluate
image quality in terms of different existing metrics would be
satisfaction and a key indicator of transmission conditions. really useful.
Meanwhile, quality is very useful in the evaluation of the The watermarking technique is one of the most promising
effectiveness or performance of image processing algorithms methods to develop Reduced or No-Reference quality metrics.
or systems. Based on the dependence on a reference image, the In this case, usually a semi-fragile watermark is embedded in the
image quality metrics can be divided into three categories: the original (or cover) image. Both the embedded watermark and
Full-Reference, Reduced-Reference and No-Reference quality the cover image will undergo the same distortion. The image
metrics. The Full-Reference quality metrics evaluate image quality can be estimated by evaluating the watermark degrada-
quality by comparing the differences between the distorted tion. The watermarking based quality estimation schemes can be
categorized into the image-feature-dependent schemes and the
image-feature-independent schemes. In the image-feature-de-
Manuscript received September 27, 2012; revised January 13, 2013; accepted pendent schemes, normally, some features extracted from the
September 17, 2013. Date of publication November 20, 2013; date of current
version January 15, 2014. The associate editor coordinating the review of this original image are embedded into the original image [9], [17],
manuscript and approving it for publication was Dr. Ton Kalker. [18]. At the receiver side, the embedded features are recon-
S. Wang and J. Zhao are with the School of Electrical Engineering and Com- structed and are used as the reference watermark. The features
puter Science, University of Ottawa, Ottawa, ON K1N 6N5, Canada (e-mail:
shawang@eecs.uottawa.ca; jyzhao@eecs.uottawa.ca). extracted from the distorted image are used as the distorted wa-
D. Zheng, W. J. Tam, and F. Speranza are with the Advanced Video Sys- termark. The watermark degradation is assessed by comparing
tems, Communication Research Centre Canada, Ottawa, ON K2H 8S2, Canada the distorted watermark with the reference watermark. Instead
(e-mail: dong.zheng@crc.ca; james.tam@crc.ca; filippo.speranza@crc.ca).
Color versions of one or more of the figures in this paper are available online
of the original image features, using the reconstructed image
at http://ieeexplore.ieee.org. features as the reference watermark may introduce additional in-
Digital Object Identifier 10.1109/TMM.2013.2291658 accuracy to the quality estimation. Moreover, for these schemes,

1520-9210 © 2013 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
312 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

it is hard to examine which kind of image features are suit- across the subbands in the wavelet decomposed image. The
able for providing accurate quality estimation. The image-fea- DWT and SPIHT together provide a good summarization of
ture-independent schemes simplify the situation [19]–[23]. In local region characteristics of an image which is important for
these schemes, the watermark is independent of the image fea- adaptive watermark embedding. In this proposed scheme, all the
tures and needs to be known at the receiver side. The water- correlated DWT coefficients across the subbands are grouped
mark degradation is evaluated using the distorted watermark together using the SPIHT tree structure. The DWT decomposed
and the original watermark. One challenging task to develop the image is further decomposed into a set of bitplane images. In this
image-feature-independent watermarking based quality metric case, each DWT coefficient is decomposed into a sequence of
is to enable the embedded watermark to accurately reflect the binary bits. The binary watermark bits are embedded into the se-
quality changes of cover image under distortions. This requires lected bitplanes of the selected DWT coefficients of the selected
the watermark being adaptively embedded in cover images ac- trees. The HVS masking is used to guide the bitplane selection.
cording to the characteristics of cover images, so that the em- As found in the experiments, the higher frequency DWT sub-
bedded watermark degrades in a similar way as the cover images bands and less significant bitplanes are more sensitive to dis-
under distortions. This is a critical part of the whole scheme and tortions, and vice versa [19]. Therefore, the robustness of the
directly affects the accuracy of quality estimation. watermark is controlled by two factors:
In [19], an image-feature-independent watermarking based (a) The percentages of the watermark bits embedded into the
quality estimation scheme was proposed which attempts to three DWT levels, respectively, and,
achieve the quality estimation accuracy of the Full-Reference (b) The selection of bitplanes for watermark embedding.
objective metrics. In the scheme, the watermark degradation Thus, for different selected trees, the watermark embedding
is measured using the True Detection Rates (TDR). Then the strengths are different.
image quality is estimated by mapping the calculated TDR to The proposed scheme is tested in terms of PSNR, wPSNR,
a quality value using an empirical mapping function which Watson JND and SSIM, and under JPEG compression,
is experimentally generated and is used as a priori at the JPEG2000 compression, Gaussian low-pass filtering and
receiver side. An iterative process is used to find the optimal Gaussian noise addition. The results show the effectiveness of
watermark embedding strength by experimentally testing the the proposed scheme.
image degradation characteristics so that the quality estimation The rest of the paper is organized as follows. Section II,
error can be minimized. This iterative process provides high Section III and Section IV present the proposed watermark
accuracy to the quality estimation. However, it also introduces embedding scheme in detail. Section V describes the watermark
relatively high computational complexity which makes it less extraction and quality evaluation scheme. Section VI details
suitable for certain applications. Meanwhile, in the scheme, the generation of the “Ideal Mapping Curve”. Section VII
the human perception characteristics have not been taken into shows the experimental results and accuracy evaluations of
consideration during the watermark embedding process and the proposed scheme. Section VIII lists the advantages of the
the quality of the watermarked image is about 40 dB in PSNR proposed scheme. Section IX concludes the paper and discusses
on average. Thus, the goal of our research in this paper is future work.
to keep the accuracy of quality estimation achieved in [19]
while improving the computational efficiency and reducing the II. THE PROPOSED WATERMARK EMBEDDING SCHEME
image quality degradation caused by the watermark embedding
The proposed watermark embedding scheme is shown in
process. Here, the term “accuracy” evaluates the correlation of
Fig. 1, which includes the watermark pre-processing, the image
the estimated quality and the quality calculated using the ex-
isting objective Full-Reference quality metrics, such as PSNR. pre-analysis, and the watermark embedding. The watermark
The closer the estimated quality to the calculated quality, the embedding process consists of the following three steps:
more accurate the quality estimation, and vice versa. (i) Apply 3-level DWT to the original image to obtain the
In this paper, we propose a new approach that well meets DWT decomposed image. The 3-level DWT decomposed
our research goal addressed above. In the proposed scheme, the subbands are denoted as shown in Fig. 2(a). These denota-
adaptive watermark embedding strength is estimated by ana- tions for the 10 DWT decomposed subbands will be used
lyzing the quality degradation characteristics of the cover image throughout the paper.
and no iterative adjustment loops are used, which significantly (ii) Embed the watermark with adaptive embedding strength
improves the computational efficiency and theoretically makes using the tree structure based watermark embedder which
the proposed scheme applicable to real-time video quality es- will be presented in Section III in detail. The output of the
timation. Moreover, the strategies including the HVS masking watermark embedder is the watermarked DWT image.
are used to guide the watermark embedding process. With the (iii) Apply 3-level inverse DWT to the watermarked DWT
proposed scheme, the quality of the watermarked images refer- image to obtain the watermarked image.
ring to the original images is about 48 dB in PSNR on average, The watermark pre-processing procedure is shown in the
which is an 8-dB improvement over the scheme in [19]. grey-dashed block with number 1 on its upper right corner. In
The proposed scheme is based on adaptive watermarking and this procedure, a two dimensional original watermark is orga-
tree structure in the DWT domain. Recently, the Set Partitioning nized column by column into a one dimensional sequence. We
in Hierarchical Trees (SPIHT) has become one of the most pop- denote the length of the original watermark sequence as len. To
ular image and video coding method because of its efficiency increase the probability of the correct watermark bit extraction
which is accomplished by exploiting the inherent similarities at the receiver side, every bit in the original watermark sequence
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 313

Fig. 1. The proposed watermark embedding process.

embedding the binary watermark bits into the selected bitplanes


of the selected coefficients. All these functions will be presented
in the following three subsections, respectively.

A. The Formation of the Tree Structure


The tree structure is formed by categorizing the DWT coef-
ficients with inherent similarities across all the DWT subbands.
The correlated coefficients build up the parent-descendants re-
lationship and form a tree [24], [25]. An example of the tree for-
mation is shown in Fig. 2(b). In the subband, 1/4 DWT co-
efficients with both odd horizontal and odd vertical coordinates
Fig. 2. Illustration of the 3-level DWT decomposed subbands (a) and the for- have no descendants. Only the coefficients with at least one even
mation of tree structure (b). coordinate have descendants and we define these coefficients
as the roots of the trees. For the DWT coefficients with odd
horizontal and even vertical coordinates, all their descendants
is repeated a few times to get a redundant watermark sequence are located in the detail subbands, where .
for watermark embedding. A parameter, , is in- For the coefficients with even horizontal and even vertical co-
troduced to express the watermark redundancy. In the proposed ordinates, all their descendants are in the detail subbands.
scheme, we set and the original watermark Correspondingly, all the descendants for the coefficients with
sequence is repeated times to get the redun- even horizontal and odd vertical coordinates are in the
dant watermark sequence with bits long. subbands. Each root has descendants in the cor-
The watermark bit extraction will be mathematically described responding detail subbands on level , where is the total
in Section V. number of DWT decomposition levels and in this paper.
The image pre-analysis procedure is shown in the Therefore, 3072 trees can be formed for a 512 512 image.
grey-dashed block with number 2 on its upper right corner. In The number of the selected trees can be evenly distributed in
this procedure, the texture characteristics and the perceptual the three orientations, , and , or more in some
masking effects of the cover image will be analyzed respec- orientation(s) and less in the other(s). With the formation of
tively in the spatial and DWT domains. Based on these image the tree structure, the locations for the watermark embedding
analysis, the adaptive watermark embedding strength is esti- can be more flexibly chosen. Moreover, the watermark bits can
mated. The details will be presented in Section IV. be embedded into the selected trees with different embedding
The position separation key is used to locate the positions for strengths.
watermark embedding.
B. The Selection of Trees and DWT Coefficients
III. TREE STRUCTURE BASED WATERMARK EMBEDDER For the applications of the watermarking based quality esti-
The tree structure based watermark embedder is designed to mation, it is desirable to embed watermark throughout the cover
embed the binary watermark bits into the selected bitplanes of image so that, even the watermarked image is locally tampered,
the selected DWT coefficients of the selected trees. In the pro- the extracted watermark can still reflect the quality degradation
posed scheme, the tree structure based watermark embedder has of the cover image. According to the length of the watermark
three functions: forming the tree structure, selecting the trees sequence, the trees for watermark embedding are chosen using
and the DWT coefficients for the watermark embedding, and the position separation key.
314 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

In the proposed scheme, to keep the embedded watermark


invisible and limit the image quality degradation caused by the
watermark embedding, the watermark bits are not embedded
into the subband of the DWT decomposed image and the
watermark bits are not embedded into the bitplanes higher than
5, where the least significant bitplane is bitplane 1. In this paper,
we denote the watermark bit assignment as ,
where , and are the number of watermark bits to be
embedded in the DWT level 1, 2 and 3 in every selected tree.
For watermark embedding, the redundant watermark sequence
is divided into segments as depicted in (1).

Fig. 3. The tree selection strategy.


(1)
where is the length of the watermark sequence. If we denote
the rows and columns of the DWT image as and , there will
be a total of trees out of trees
selected for the watermarking process. Here, we define another
parameter to control the percentage
of trees selected from the three orientations. With , we can
embed the watermark into the three orientations evenly or by
focusing more on one or two orientations.
To minimize the quality degradation of the cover image
caused by the watermark embedding, two strategies are used
for the tree selection:
(a) The trees selected from the three DWT orientations are
non-overlapping in position.
For example, in Fig. 3(a), we use the square blocks in
, and to indicate the tree positions avail- Fig. 4. The tree selection from the three DWT orientations. (a) Graphical il-
able in the three orientations for selection. The trees , lustration. (b) Experimental illustration.
and shown in Fig. 3(a) are all carrying the spatial
information of block 0 in . Thus, if all the trees A,
B and C are embedded with watermark bits, the quality selected trees segment by segment following the order of
of the top left corner of the image will degrade severely. the selected tree position.
Therefore, if tree is chosen for the watermark embed- After the tree selection, we start to embed the watermark bit
ding, trees B and C will not be selected. In this case, segments into the selected trees referring to . On the th
for an image with trees, there DWT level, the watermark bits are embedded into the DWT
are non-over- coefficients one by one until the number of is reached,
lapping tree positions available for the tree selection. In where .
Fig. 3(b), the ‘X’ marked positions represent the trees se- An experimental example of the tree and DWT coefficients
lected from the three DWT orientations for the watermark selection is shown in Fig. 4(b). All the selected DWT coeffi-
embedding. The ‘ ’ marked positions represent the sep- cients are marked as dark points. The darker points means that
aration between any two selected tree positions. In our the lower significant bitplanes are selected, and vice versa. The
implementation, we use uniform separation, defined as bitplane selection will be presented in Section IV-B. In Fig. 4(b),
. will be output as the the position separation key is 3.
position separation key and will be transmitted to the re-
ceiver side. Here, we set .
C. The Watermark Embedding
(b) The trees are selected throughout the DWT decomposed
image. The binary watermark bits are embedded into the selected
The selected trees are further distributed into the bitplanes of the selected DWT coefficients. Here, we denote the
three orientations, as described by . In the proposed watermark bit as , the DWT coefficient bit on the selected
scheme, we set , which means that bitplane as and the watermarked DWT coefficient bit as .
the trees are evenly selected from the three orientations Then, the watermark bit will be embedded using the following
as illustrated in Fig. 4(a). In this figure, the selected trees equation:
are numerically ordered as . As mentioned
previously, the watermark sequence is divided into if
(2)
segments. Thus, the watermark will be embedded into the if
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 315

IV. THE ESTIMATION OF ADAPTIVE WATERMARK


EMBEDDING STRENGTH

As mentioned in Section I, the watermark embedding strength


is controlled by the empirical watermark bit assignment and the
bitplane selection, which will be presented in the following two
subsections, respectively.

A. The Empirical Watermark Bit Assignment


The watermark bit assignment is denoted as
. Based on the analysis on image content com-
plexity, we empirically assign watermark bits to the 3 DWT
levels in the selected tree using the following steps:
(a) Analyze the content complexity of the cover image and
calculate a complexity index.
(b) Categorize the test images into different groups according
to their complexity indices.
(c) Assign watermark bits to the 3 DWT levels of the cover
image.
The watermark bit assignment will be the same for all the
selected trees in one image and may be different for different
images.
1) The Analysis of Image Content Complexity: The quad-tree
decomposition based complexity analysis is used in the pro-
posed scheme for a better match with the DWT [26]. For gray
scale images, the intensity difference, , is used to verify
Fig. 5. Illustration of the image content complexity analysis. (a)&(c)&(e) Orig-
whether a further quad-tree decomposition is needed. Here, we inal images. (b)&(d)&(f) Quad-tree decomposed images. (b):
define an intensity difference threshold as . If , . (d): . (f): .
the image or current block will be decomposed into 4 sub-blocks
until or the size of the sub-block reaches 1. Each
quad-tree decomposition is recorded as a decomposition node. Three experimental examples are shown in Fig. 5. These
The depth of the decomposition is denoted as the level of de- quad-tree decomposed images are achieved using the threshold
composition. The content complexity of the cover image is as- , where the maximum intensity value of the
sessed using the following equation [26]: cover image is not bigger than 1. In this case, the brighter the
quad-tree decomposed image, the more complex the cover
(3) image. The complexity indices are listed with the quad-tree
decomposed images.
2) The Watermark Bits Assignment: With the complexity in-
where is the current quad-tree decomposition level; is the
dices, the watermark bits are empirically assigned to the images
highest decomposition level; is the number of quad-tree de-
using the following steps:
composition nodes on level . Then, the calculated complexity
(a) The complexity indices are divided into 6 groups using
values of all the images in our image library are normalized. In
(4). One integer index is associated with one group.
the proposed scheme, the normalized complexity value is used
as the complexity index which locates in [0, 1].
The image content complexity analysis evaluates how much
details that an image contains. A higher complexity value in-
dicates that the image is more complex and the image contains (4)
more detail information. Comparing to a less complex image,
the quality of a more complex image degrades faster against the
same distortion [19]. For this case, to reflect the quality degra-
dation of the cover image, we need to embed more watermark where is the group index; is the complexity
bits into the lower DWT levels of a more complex image. For index. , , , , and are the empirical grouping
a less complex image, we consider to embed more watermark thresholds. These thresholds may be different for different
bits into the higher DWT levels. distortions.
316 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

(b) With the group indices, the watermark bits are assigned to where
the images using (5).
if
o.w.
if
if
(5) if
for HL blocks
for HH blocks
for LH blocks
and
(b) Background luminance: Intensity variations are less vis-
ible over the brighter and darker areas. The luminance
when
masking is denoted as .
when (6)
when

Recall that, in (1), trees are if


(8)
selected for the watermark embedding, where is o.w.
the length of the redundant watermark sequence. Thus,
(c) Spatial masking or edge proximity: The human eyes are
for images with different complexities, the number of se- more sensitive to noise addition near edges or contours of
lected trees, , and the position separation key, , images. This factor, , is evaluated using the empiri-
may be different. cally scaled local energy of the DWT coefficients in all
detail subbands.

B. The HVS and Bitplane Based Perceptual Masking

The other factor that controls the watermark embedding (9)


strength is the bitplane selection for the watermark embedding. where is a weighting parameter and the suggested value
By decomposing the DWT image into a set of binary bitplanes, for is presented in the following equation [27].
each DWT coefficient is decomposed into a binary bit sequence. if
The less significant bits are more sensitive to distortions than (10)
o.w.
the more significant bits. If we embed the whole watermark
in the least significant bitplane of the DWT image, we obtain (d) Texture sensitivity: Intensity variations in highly textured
the most fragile watermark and the least quality degradation areas are less visible than those in the flat-field areas of
to the cover image, and vice versa. In the proposed scheme, images. This masking factor, , is estimated using the
the bitplane selection is guided by the analysis of the HVS local variance of the corresponding DWT coefficients in
masking effects on the DWT decomposed image. The analysis the LL subband.
of the HVS masking effects helps to increase the capacity for
watermarking without causing more quality degradation to the
cover image.
(11)
The bitplane selection includes two steps: calculating the
where and .
HVS masks and mapping the calculated HVS masks to bit-
The HVS mask is achieved by polling the four factors listed
plane indices. The achieved bitplane indices decide which
above and is computed using (12).
bitplanes of the selected DWT coefficients are used to embed
the watermark bits. The HVS masking calculation and the (12)
HVS-to-bitplane mapping are presented in the following two
subsections. where denotes the HVS mask; is a scaling parameter.
1) The HVS Masking: The HVS masking presented in [27], As analyzed in [27], the suggested value for is 1/2, which
[28] is used in the proposed scheme. Accordingly, four factors implies that intensity variations having values lower than half
greatly affect the behavior of the HVS: of are assumed invisible. The suggested
(a) Band sensitivity or frequency masking: Intensity varia- value for and is 0.2 [27]. In (7) through (12), is
tions are less visible in high resolution subbands and are the luminance value of the pixel, , in one detail subband
also less visible in the diagonally decomposed blocks, on level ; indicates the current DWT
. This factor is expressed using (7). level in the HVS masking calculation and is the max-
imum level applied in the DWT decomposition in this paper;
(7) is the luminance value of pixel
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 317

Fig. 6. Illustration of the generated HVS masks. (a) The original image Bar-
bara. (b) The nine HVS masks generated for the nine DWT detail blocks.

in the subband which corresponds


to pixel in a detail subband on level .
As proposed in Section III-B, the binary watermark bits are
not embedded in the subband so that the invisibility of the
embedded watermark can be further improved. Therefore, the Fig. 7. Illustration of the thresholds selection. (a)&(d): Original image Bar-
HVS masks are only calculated for the nine detail DWT sub- bara and Treefrog. (b)&(e): Normalized HVS mask of DWT subband HL1 of
(a)&(d). (c)&(f): Histogram of the mask shown in (b)&(e).
bands, with one mask for each. The generated HVS masks for
image Barbara is shown in Fig. 6(b).
2) Mapping From HVS Mask to Bitplane Indices: To
Therefore, each DWT coefficient in the selected trees has its
use the HVS mask in the watermark embedding process, a
own . At the receiver side, the strategy presented
mapping relationship from the coefficients of the HVS mask
above will also be applied on the distorted image to locate the
to the bitplane indices is experimentally defined using the
bitplanes for the watermark extraction. To increase the proba-
multiple-band-thresholding method. Considering that different
bility of correct extraction of the watermark bits, all the calcu-
HVS masks have different distributions, the thresholds used in
lated values on the DWT level in each tree are
the mapping procedure should be able to change with the shape
averaged. In other words, in each selected tree, for all the se-
of the distribution of the HVS mask. To further limit the quality
lected coefficients, the watermark bits will be embedded on the
degradation caused by the watermarking process, only the
same bitplane. Thus, the bitplane indices, , are updated using
bitplanes from 1 to 5 are used for the watermark embedding.
the following equation:
The thresholds calculation and the mapping procedure are
listed as follows:
(a) The HVS mask is first normalized using its maximum co-
efficient so that the mapping relationship can be consistent
with different HVS masks.
(b) The coefficients of the HVS masks are sorted in an as-
cending order and the sorted sequence is denoted as Sort-
where is the averaged bitplane of the DWT
mask.
coefficients located in a specific tree on level and orientation
(c) The thresholds are calculated using (13).
.

(13)

where , is the total number of the


coefficients of one HVS mask, the denominator indicates
that 5 bitplanes are used for the watermark embedding.
Two examples of the thresholds selection are shown in
(d) The mapping is done using (14).
Fig. 7. Fig. 7(b) and (e) are the HVS masks respectively
calculated on the DWT subband HL1 of image Barbara and
image Treefrog. Thus, the size of these two HVS masks
are 1/4 of the image size. Then the HVS-to-bitplane map-
ping thresholds are calculated using (13). Fig. 7(c) and (f)
graphically illustrate the thresholds selection. The com-
(14) puted thresholds for Fig. 7(c) are
where are the coordinates of a selected DWT coeffi- . The thresholds for Fig. 7(f)
cient. means the bitplane index achieved for are .
the pixel located at on DWT level with orientation With the calculated thresholds, the coefficients of the HVS
. is the value of the HVS mask coefficient. mask which correspond to the selected DWT coefficients are
318 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

Fig. 8. Watermark extraction and quality estimation.

mapped to the bitplane indices. In this way, the bitplanes of the


selected coefficients for the watermark embedding are located.

V. THE PROPOSED WATERMARK EXTRACTION


AND QUALITY ESTIMATION SCHEME

The proposed watermark extraction and quality estimation


scheme is shown in Fig. 8, where the position separation key
is used to locate the watermarked DWT coefficients.
The watermark bit assignment is retrieved using the image
group index transmitted from the sender side. The bitplane in-
dices for watermark extraction are obtained by calculating the
HVS masks of the distorted watermarked image. As presented in
Section IV-B-2, in one tree, the bitplane indices for all the DWT
coefficients on each DWT level are averaged. This strategy ef-
fectively reduces the watermark extraction error caused by the
bitplane selection in the watermark extraction scheme.
Recall that . The extracted redundant wa-
termark sequence is used to recover the three distorted water-
marks. Then, the three distorted watermarks are compared bit
by bit and the watermark is extracted using (15).

(15) Fig. 9. The “ideal mapping curve” generator.

where is the extracted watermark bit with coordinates


; is the number of extracted 1 s and is the number VI. THE GENERATION OF “IDEAL MAPPING CURVE”
of extracted 0 s. The “Ideal Mapping Curve” experimentally defines the re-
Then, the extracted watermark is compared with the original lationship between the degradation of the watermark and the
watermark bit by bit and the True Detection Rates (TDR) is quality of the distorted image. The “Ideal Mapping Curve” is
calculated using (16). obtained from a set of TDR-Quality curves and is used to map
the calculated TDR to a possible quality value which can be
in terms of the existing Full-Reference quality metric, such as
(16) PSNR, wPSNR, Watson JND or SSIM.
As described in (17), the image quality is estimated by map- For the image quality estimation in terms of one Full-Ref-
ping the calculated TDR to a quality value by referring to a map- erence quality metric under one type of distortion, one “Ideal
ping function. Mapping Curve” needs to be generated. As shown in Fig. 9, the
“Ideal Mapping Curve” is generated by locally averaging the
(17) TDR-Quality curves obtained by testing a number of images.
An example of “Ideal Mapping Curves” generated respec-
where is the estimated quality; is the mapping function, tively using 10, 20, 50 and 100 different natural images in
which is the “Ideal Mapping Curve” in the proposed scheme. terms of JND under Gaussian low-pass filtering are shown in
The generation of the “Ideal Mapping Curve” will be presented Fig. 10. From this figure, we can see that the four generated
in the next section. “Ideal Mapping Curves” are nearly identical. As mentioned
When the calculated TDR is mapped onto the “Ideal Map- previously, the adaptive watermark embedding strength is
ping Curve”, it could possibly lie between two neighboring TDR assigned by analyzing the degradation characteristics of the
values on the curve. In this case, linear interpolation is used to cover image. With the adaptive watermark embedding strength,
estimate the image quality. we expect the TDR-Quality curves tested using a number of
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 319

Fig. 11. Examples of the generation of “ideal mapping curves” under Gaussian
low-pass filtering.

The same averaging rule is used for the “Ideal Mapping


Curve” generation under other distortions. Once the “Ideal
Mapping Curve” is generated, it is stored at the receiver side
Fig. 10. “ideal mapping curves” generated using different number of test and no transmission is needed.
images.
VII. EXPERIMENTAL RESULTS AND ACCURACY EVALUATION
images as convergent as possible. The more convergent the The data need to be transmitted to the receiver side are:
tested TDR-Quality curves, the more accurate the quality es- (a) The original binary watermark which is 48 48 pixels
timation. The similar convergence of the TDR-Quality curves (2304 bits). It is unnecessary to transmit the watermark
obtained respectively using 10, 20, 50 and 100 images results if both the sender and receiver sides use the same water-
in the four nearly-identical “Ideal Mapping Curves” as shown mark.
in Fig. 10, which also indicates the flexibility of choosing the (b) The position separation key.
number of images for the “Ideal Mapping Curve” generation. (c) The empirical group index for the watermark bit assign-
With the proposed scheme, we suggest to use at least 10 natural ment.
images for generating the “Ideal Mapping Curves” to make
it statistically meaningful. In our implementation, we use 50 A. The Original Images and the Original Watermark
natural images selected from our image library to generate the There are 150 gray images in our image library including all
“Ideal Mapping Curves”. The scope of our image library is the reference images from Cornell A57 image database, IVC
presented in Section VII-A. image database, TID2008 image database, and CSIQ image
Two examples of the “Ideal Mapping Curve” gener- database. Besides these image databases, our image library also
ation under Gaussian low-pass filtering are shown in includes computer generated images and more natural images.
Fig. 11. The gray dots are the TDR-Quality curves cal- All of these images are 512 512 in size and they contain
culated using the 50 test images. The 50 TDR-Quality different textures, such as portraits, plants, animals, sceneries,
curves are computed when the filter standard deviation
buildings, crowd, animation and computer generated patterns.
.
The first 50 images in the image library are used to generate the
When , the PSNR of the low-pass filtered image
“Ideal Mapping Curves” and other empirical parameters. To
is about 25 dB on average. The “Ideal Mapping Curve” is
avoid losing generality, the 51st to the 150th images are used
denoted using black dot dashed curve and is calculated by
to evaluate the performance of the proposed scheme.
locally averaging the 50 gray dot curves. The local averaging
The original binary watermark used in the experiments is 48
consists of the following steps:
48 in size and is randomly generated. The watermark can be
(a) Low-pass filter the test images respectively using the
changed to any size or pattern.
filter standard deviations listed above. Calculate the
quality values and TDR values under each . Find Fig. 12 shows the watermarked images of
the possible quality range, which is [0, 25] JND for Fig. 5(a), (c) and (e), respectively. The watermark bit
Fig. 11(a) and [20, 49] dB for Fig. 11(b). assignment, , associated with the images are used for
(b) List all the desired target quality values in the quality watermark embedding under Gaussian low-pass filtering. The
range found in step (a), which is [0.02, 0.1:0.1:1, 2:1.5:14, quality of the three example watermarked images shown in
16:2:25] JND for Fig. 11(a) and [49:-0.5:48, 47,46:-1:35 Fig. 12 are more than 47.5 dB in PSNR, which indicates that
34:-2:20] dB for Fig. 11(b). the watermark embedding process does not have significant
(c) Compare each calculated quality with all the target quality impact on the image quality.
values. By finding the closest target quality, all the calcu-
B. Summary of the Experiments
lated quality values and their corresponding TDR values
are categorized into different groups. The proposed scheme is evaluated under JPEG compres-
(d) Average the TDR values in each group. sion, JPEG2000 compression, Gaussian low-pass filtering and
(e) Generate the “Ideal Mapping Curve” using the target Gaussian noise addition. Under each distortion, a watermarked
quality values and the averaged TDR values. image is distorted using a variety of distortion strengths and
320 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

TABLE I
RANGE OF DISTORTION STRENGTHS AND CORRESPONDING
WATERMARKED IMAGE QUALITIES IN PSNR

In Table I, , and respectively stand for


, and . For JPEG, the distortion
strength is the compression strength and is denoted using
“Quality Factor” in Matlab. For JPEG2000, the distortion
strength is the compression rate. For Gaussian low-pass fil-
tering (denoted as LPF in the table), the distortion strength is
controlled by the standard deviation of the low-pass filter. For
Gaussian additive noise, the distortion strength is the standard
deviation of the Gaussian noise.
Fig. 12. Illustration of the effects of watermark embedding. The accuracy of the quality estimation under both and
(a) ; Watermark bits assignment: ; will be evaluated and shown in Section VII-E. Note that
. (b) ; Watermark bits assign-
ment: ; . (c) ;
is an averaged quality value. It is normal if some quality
Watermark bits assignment: ; . points smaller than appear in the accuracy evaluation
figures.

the quality of the distorted images is estimated in terms of D. The Empirical Grouping Thresholds
PSNR, wPSNR, Watson JND [29] and SSIM [2], respectively. For each set of experiments, the grouping thresholds used in
Therefore, totally 16 sets of experiments are conducted. In (4) are empirically set by testing the first 50 images in the image
each set of the experiments, 100 images are tested. To assess library using the following steps:
the performance of the proposed scheme, the quality of the (a) Based on the research done in [19], empirically set up the
distorted image comparing to the original image is calculated watermark bit assignment, , as listed in (5).
as well. With both the estimated quality and calculated quality, (b) With each , test the quality estimation error using the
the quality estimation accuracy is evaluated using MAE (Mean 50 images.
Absolute Error), Pearson correlation coefficient, and RMSE (c) For each test image, find the minimum quality estimation
(Root Mean Squared Error) respectively. In the experiments, error and the corresponding optimal .
the image quality degradation caused by both the watermark (d) Calculate the content complexity values for the 50 test
embedding and the distortion is considered in the performance images using the local intensity difference threshold 0.17,
evaluation. where the maximum intensity value for the images is 1.
(e) Referring to the optimal assigned to the 50 test im-
C. Choosing the Range for the Distortion Strength ages, categorize the calculated content complexity values
The watermarked images are distorted using a set of dis- into 6 groups.
tortion strengths varying in a selected range, , (f) Set up the empirical thresholds by examining the catego-
where is the distortion strength; and respectively are rized content complexity values.
the lowest and the highest distortion strength used in the ex- For different set of experiments, the empirical grouping
periments. We choose the lowest distortion strength available thresholds may be different. The thresholds used in the experi-
in Matlab as . For example, under JPEG compression, the ments are listed in Table II. As found in the experiments, under
quality factor controls the distortion strength and . Gaussian noise distortion with a fixed noise standard deviation,
is so set that the average quality of the watermarked images dis- the quality of different distorted images varies slightly no
torted using , , is in the range of [24.50, 25.50] dB matter how complex the images are. Thus, under Gaussian
in PSNR. , where the noise distortion, we use the grouping thresholds the same as
is the watermarked image and is the original image. On those used under JPEG compression for convenience. These
the other hand, when the quality of the distorted image is below thresholds are stored at the receiver side once they are tested.
30 dB in PSNR, the image quality is poor. The lower the image
quality, the more meaningless to focus on the quality estimation E. Experimental Results Under Different Distortions
accuracy. Thus, we choose another high boundary for : For the four types of distortions, the distortion strengths used
if in PSNR. The 150 images in in the experiments are listed in the following:
the image library are used to choose , and . The range • JPEG: ;
of the distortion strengths for the four types of distortions used • JPEG2000:
in the experiments are listed in the following table: .
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 321

Fig. 13. The generated “ideal mapping curves” in terms of different full-reference quality metrics under different distortions.

TABLE II
THE EMPIRICAL THRESHOLDS USED FOR THE 16 SETS OF EXPERIMENTS , the JND value increases rapidly with the decreasing TDR
value. When TDR decreases from 0.4 to 0.31, the JND value in-
creases from 8 to 40. Thus, we need to use a TDR value varying
in a 0.09 value range to estimate image quality varying in a 32
JND value range. In this case, a small variation of the TDR value
may result in a relatively significant quality estimation error. So,
• Low-pass filtering: we can predict that this rapid increase of JND will greatly reduce
. the accuracy of the quality estimation when . For
• Gaussian noise distortion: additional observations, the quality estimation accuracy when
. the calculated quality is smaller than 8 JND is evaluated as well.
For the 16 sets of experiments, the 16 generated “Ideal Similar analysis can be made to Fig. 13(g), (k), and (o).
Mapping Curves” are shown in Fig. 13(a)–(p). For example, Fig. 13(d), (h), (l) and (p) are the “Ideal Mapping Curves”
Fig. 13(a), (e), (i) and (m) show the “Ideal Mapping Curves” used to estimate image quality in terms of SSIM. In Fig. 13(d),
for the quality estimation in terms of PSNR. With increasing the left half of the curve has a very big slope. When the TDR
distortion strength, the calculated PSNR value and the calcu- decreases from 0.42 to 0.31, the calculated SSIM changes from
lated TDR value decrease monotonically. 0.9871 to 0.7. The SSIM values drop rapidly corresponding a
Fig. 13(c), (g), (k) and (o) are the “Ideal Mapping Curves” small change of TDR value. It is straightforward that a small
used to estimate image quality in terms of JND. With increasing variation of the TDR value may cause relatively bigger estima-
distortion strength, the calculated TDR decreases and the cal- tion error and the divergence of the quality points referring to
culated JND value increases. When TDR equals to 1, the JND the solid matching line may be magnified.
is 0. In Fig. 13(c), is a slope change point and The 16 sets of experimental results of image quality estima-
JND, where is the quality. When tion tested with distortion strengths are shown in
322 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

Fig. 14. The experimental results tested with distortions in terms of different Full-Reference quality metrics under different distortions.

Fig. 14(a)–(p). In the figures, the x-axis is the estimated quality. scattering. The poorer the image quality, the more meaning-
The y-axis is the correspondingly calculated quality. The black less to focus on the quality estimation accuracy. The accuracy
points are the quality points achieved under different distortion of the image quality estimation evaluated using MAE, Pearson
strengths by testing the 100 images. The solid line is the quality correlation coefficient and RMSE is summarized in
matching line with a slope of 1, which indicates that the esti- Table III, Table IV and Table V, respectively. In the tables,
mated quality equals to the calculated quality. The discrepancy and respectively indicate the range of the distortion strengths
from the solid line to the quality points indicate the accuracy of and , which are described in Section VII-C and
the quality estimation. The more convergent the quality points represents stronger distortion strength than . LPF stands
to the solid line, the more accurate the quality estimation and for low passing filtering, and Gaussian stands for Gaussian noise
vice versa. If all the quality points fall on the solid line, the es- addition.
timated quality equals to the calculated quality and the highest From Table III through Table V, we can see that:
quality estimation accuracy is achieved. For each sub-figure in (a) The proposed scheme has good accuracy to estimate
Fig. 14, the corresponding quality estimation accuracy in MAE image quality in terms of PSNR and wPSNR. Under
is listed. Recall that in PSNR. When the the four mentioned distortions, with both the distortion
quality of a distorted image is lower than 30 dB in PSNR, the strengths and , the calculated for
image quality becomes poor and the quality points become more the quality estimation in PSNR are respectively lower
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 323

TABLE III TABLE VI


SUMMARY OF THE ACCURACY IN MAE COMPARISON BETWEEN THE ACCURACY OF OUR PREVIOUS
SCHEME AND THE PROPOSED SCHEME UNDER JPEG COMPRESSION

TABLE IV
SUMMARY OF THE ACCURACY IN PEARSON CORRELATION In the proposed scheme, four sets of parameters and thresh-
olds have been introduced: the grouping thresholds for calcu-
lating in (4), the watermark bit assignment parameters
in (5), the parameters used to calculate the HVS mask
in (12), and the mapping parameters from HVS masks to
bitplane indices in (14). The first set of parameters in (4) is
image independent and distortion dependent. The rest three sets
of empirical parameters are image and distortion independent,
TABLE V which means that they are fixed for all images and all types of
SUMMARY OF THE ACCURACY IN RMSE
distortions. All these parameters and thresholds work together
to balance the robustness of algorithm and the quality degrada-
tion caused by the watermark embedding.

VIII. ADVANTAGES OF THE PROPOSED SCHEME


Based on the performance evaluation in Section VII-E, the
advantages of the proposed adaptive watermarking and tree
structure based quality metric are listed in the following:
than 0.74 dB and 0.8710 dB; and the calculated (a) Computational efficiency. We have mentioned that our
for the quality estimation in wPSNR are respectively previous work [19] requires an average of 12 recursive
lower than 0.77 dB and 0.9405 dB. The gives loops to find the optimal watermark embedding strengths
similar results as the does. The calculated Pearson for the test images. With the proposed scheme, the exten-
correlations are all higher than 0.977 which indicates the sive iterative process is not needed and the computational
high correlations between the estimated quality and the efficiency is improved about 17 times over that of our pre-
calculated quality. vious work [19] based on the computation cycle counting
(b) To estimate image quality in terms of JND, the proposed using profiling tools.
scheme has the highest accuracy under Gaussian low-pass (b) The quality loss caused by the watermark embedding is
filtering with the distortion strength in both the range very small. For 150 images, the PSNR of the watermarked
and . Under other distortions, the proposed image is 48.1476 dB on average.
scheme has good quality estimation accuracy with (c) The proposed scheme can assess image quality in terms
and acceptable quality estimation accuracy with of PSNR, wPSNR, Watson JND and SSIM with good
. accuracy. Moreover, it can be used to evaluate image
(c) To estimate image quality in terms of SSIM, the proposed quality degradation under different distortions, such as
scheme has good quality estimation accuracy under all JPEG compression, JPEG2000 compression, Gaussian
the four distortions with and acceptable low-pass filtering and Gaussian noise addition. The pro-
accuracy with . posed scheme can be developed to estimate image quality
Under JPEG compression, the performance of the proposed in terms of other Full-Reference quality metrics and or
scheme is compared with that of the scheme in [19]. Our under other distortions.
previous scheme [19] was tested by estimating image quality (d) It works as a Reduced or No-Reference quality evaluation
in terms of PSNR, wPSNR, JND under JPEG compression and scheme which is more desirable in communication sys-
achieved good accuracy. The quality factors [100:-5:20] are tems. With the proposed scheme, the cover image is un-
used in [19] for the image quality estimation. The accuracy necessary to be transmitted to the receiver side for quality
of the image quality estimation of both our previous scheme evaluation, which greatly saves the transmission band-
and the currently proposed tree structure based scheme under width.
JPEG compression with quality factors [100:-5:20] are listed
in Table VI for comparisons. From both the calculated MAE IX. CONCLUSIONS AND FUTURE WORK
and the calculated Pearson correlations, , we can see that In this paper, a new watermarking based quality estimation
the two schemes can provide comparable accuracy for image scheme is presented. The proposed scheme is designed to es-
quality estimation. timate image quality in terms of the existing Full-Reference
324 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 16, NO. 2, FEBRUARY 2014

quality metrics, such as PSNR, wPSNR, JND and SSIM. Thus, [15] H. Wu and M. Yuen, “A generalized block edge impairment metric for
at the receiver side of a communication system, without the orig- video coding,” Signal Process. Lett., vol. 4, no. 11, pp. 317–320, 1997.
[16] F. Battisti, M. Carli, and A. Neri, “No-reference quality metric for color
inal image, the quality of a distorted image can still be assessed. video communication,” in Proc. Int. Workshop Video Processing and
Based on the tree structure, the binary watermark is em- Quality Metrics, 2012, pp. 1–6.
bedded into the selected bitplanes of the selected DWT [17] A. Bhattacharya, S. Palit, N. Chatterjee, and G. Roy, “Blind assessment
of image quality employing fragile watermarking,” in Proc. IEEE 7th
coefficients with adaptive watermark embedding strength. Int. Symp. ISPA, 2011, pp. 431–436.
The watermark embedding strength is assigned to an image [18] S. Altous, M. K. Samee, and J. Gotze, “Reduced reference image
by pre-analyzing its content complexity in the spatial domain quality assessment for jpeg distortion,” in Proc. IEEE 7th Int. Symp.
ELMAR, 2011, pp. 97–100.
and the perceptual masking effect of the DWT decomposed [19] S. Wang, D. Zheng, J. Zhao, W. J. Tam, and F. Speranza, “An image
image in the DWT domain. Meanwhile, the watermark is not quality evaluation method based on digital watermarking,” IEEE
embedded in the approximation subband, which reduces loss Trans. Circuits Syst. Video Technol., vol. 17, no. 1, pp. 98–105, 2007.
[20] D. Zheng, J. Zhao, W. Tam, and F. Speranza, “Image quality mea-
in image quality caused by embedding the watermark. The surement by using digital watermarking,” in Proc. IEEE Int. Workshop
experimental results show that the proposed scheme works Haptic, Audio and Visual Environments and Their Applications, 2003,
effectively. pp. 65–70.
[21] M. Farias, M. Carli, and S. Mitra, “Objective video quality metric based
In future work, the proposed scheme will be further devel- on data hiding,” IEEE Trans. Consum. Electron., vol. 51, no. 3, pp.
oped to estimate the quality of an image distorted by multiple 983–992, 2005.
distortions. Meanwhile, experiments about image quality esti- [22] E. Nezhadarya, Z. Wang, and R. Ward, “Image quality monitoring
using spread spectrum watermarking,” in Proc. IEEE Int. Conf. Image
mation in terms of subjective quality scores will be conducted. Processing, 2009, pp. 2233–2236.
Since the proposed scheme has good computational efficiency, [23] A. Mishra, A. Jain, M. Narwaria, and C. Agarwal, “An experimental
it is feasible to further develop the proposed scheme for video study into objective quality assessment of watermarked images,” Int.
J. Image Process., vol. 5, no. 2, pp. 199–219, 2011.
quality evaluation. [24] D. Taubman and M. W. Marcellin, JPEG2000: Image Compres-
sion Fundamentals, Standards and Practice. Norwell, MA, USA:
Kluwer, 2002.
[25] A. Said and W. Pearlman, “A new, fast and efficient image codec based
on set partitioning in hierarchical trees,” IEEE Trans. Circuits Syst.
REFERENCES Video Technol., vol. 6, no. 3, pp. 243–250, 1996.
[26] M. Jamzad and F. Yaghmaee, “Achieving higher stability in water-
[1] I. J. Cox, M. L. Miller, J. A. Bloom, J. Fridrich, and T. Kalker, Digital marking according to image complexity,” Scientia Iranica J., vol. 13,
Watermarking and Steganography. San Mateo, CA, USA: Morgan no. 4, pp. 404–412, 2006.
Kaufmann, 2008. [27] M. Barni, “Improved wavelet-based watermarking through pixel-wise
[2] Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli, “Image masking,” IEEE Trans. Image Process., vol. 10, no. 5, pp. 783–791,
quality assessment: From error visibility to structural similarity,” IEEE 2001.
Trans. Image Process., vol. 13, no. 4, pp. 600–612, 2004. [28] A. Lewis and G. Knowles, “Image compression using the 2-D wavelet
[3] S. Wolf, M. Pinson, S. Voran, and A. Webster, “Objective quality as- transform,” IEEE Trans. Image Process., vol. 1, no. 52, pp. 244–250,
sessment of digitally transmitted video,” in Proc. IEEE Pacific Rim 1992.
Conf. Communications, Computers and Signal Processing, 1991, vol. [29] [Online]. Available: http://vision.arc.nasa.gov/dctune/(last visited in
2, pp. 477–482. Sep. 2011).
[4] A. Watson, Q. Hu, and J. McGowan, “Digital video quality metric [30] D. Sheskin, Handbook of Parametric and Nonparametric Statistical
based on human vision,” J. Electron. Imag., vol. 10, pp. 20–29, 2001. Procedures, 2nd ed. London, U.K.: Chapman & Hall CRC, 2000.
[5] A. Shnayderman, A. Gusev, and A. Eskicioglu, “An SVD-based
grayscale image quality measure for local and global assessment,”
IEEE Trans. Image Process., vol. 15, no. 2, pp. 422–429, 2006.
[6] G. Zhai, W. Zhang, X. Yang, S. Yao, and Y. Xu, “GES: A new image
quality assessment metric based on energy features in Gabor transform
domain,” in Proc. IEEE Int. Symp. Circuits and Systems, 2006, pp. Sha Wang received her B.Eng in thermal power
1715–1718. engineering from North China Electric Power Uni-
[7] M. Carnec, P. Callet, and D. Barba, “Full reference and reduced ref- versity, China in 2002, and received her Master’s
erence metrics for image quality assessment,” in Proc. 7th Int. Symp. and Ph.D. degree in electrical engineering from
Signal Processing and Its Applications, 2003, vol. 1, pp. 477–480. University of Ottawa, Canada, respectively in 2005
[8] Q. Li and Z. Wang, “Reduced-reference image quality assessment and 2013. Her research interests are on image and
using divisive normalization-based image representation,” IEEE J. video processing.
Select. Topics Signal Process., vol. 3, no. 2, pp. 201–211, 2009.
[9] Z. Wang, G. Wu, H. R. Sheikh, E. P. Simoncelli, E. H. Yang, and A. C.
Bovik, “Quality-aware images,” IEEE Trans. Image Process., vol. 15,
no. 6, pp. 1680–1689, 2006.
[10] T. Kusuma, H. Zepernick, and M. Caldera, “On the development of a
reduced-reference perceptual image quality metric,” in Proc. Systems
Communications, 2005, pp. 178–184.
[11] P. Lecallet, C. Viard-Gaudin, S. Pechard, and E. Caillault, “No refer- Dong Zheng received his B.E. degree in electrical
ence and reduced reference video quality metrics for end to end QoS engineering from Wuhan University, China in 1996
monitoring,” IEICE Trans. Commun., vol. E85-B, no. 2, pp. 289–296, and received his M.A.Sc and Ph.D. in electrical engi-
2006. neering from University of Ottawa, Canada respec-
[12] D. Turaga, Y. Chen, and J. Caviedes, “No reference PSNR estimation tively in 2003 and 2008. He worked as a senior re-
for compressed pictures,” Signal Process.: Image Commun., vol. 19, search engineer in Dilithiumnetwork Inc., California
no. 2, pp. 173–184, 2004. from 2007 to 2008. He worked as a research scientist
[13] P. Marziliano, F. Dufaux, S. Winkler, and T. Ebrahimi, “A no-reference at Communication Research Centre of Canada from
perceptual blur metric,” in Proc. IEEE Int. Conf. Image Processing, 2008 to 2012. Since 2012, he has been working as
2002, vol. 3, pp. 57–60. a staff software engineer of Magnum Semiconductor
[14] W. Ci, H. Dong, Z. Wu, and Y. Tan, “Example-based objective quality on video processing. His current research interests in-
estimation for compressed images,” IEEE Multimedia, to be published. clude digital video and image coding and quality assessment.
WANG et al.: ADAPTIVE WATERMARKING AND TREE STRUCTURE BASED IMAGE QUALITY ESTIMATION 325

Jiying Zhao (M’00) received the Ph.D. degree Filippo Speranza received his Ph.D. in Experi-
in electrical engineering from North China Elec- mental Psychology from the University of Toronto
tric Power University, and the Ph.D. degree (Canada) in 1999. From 2001 to 2013, he worked
in computer engineering from Keio University. as Research Scientist with Industry Canada—Com-
He is a professor with the School of Electrical munications Research Centre (CRC), where he
Engineering and Computer Science, University of investigated stereoscopic imaging and subjective
Ottawa, Ottawa, Canada. His research interests picture quality assessment techniques and methods.
include image and video processing and multi- He is an adjunct professor at the University of
media communications. Dr. Zhao is a member Ottawa.
of the Institute of Electronics, Information and
Communication Engineers (IEICE) and is also a
member of the Professional Engineers Ontario (PEO).

Wa James Tam received a Ph.D. degree in experi-


mental psychology from York University in Toronto,
Canada in 1990. His graduate thesis was on the
effect of visual fixation on binocular eye movement
latencies. He embarked on his first project, on video
coding based on viewer’s eye movement patterns,
at the Communications Research Center (CRC) in
Ottawa, Canada in 1989. From then on he became
a senior research scientist and a project leader on
3D with the Advanced Video Systems Group at
CRC. After 23 years at CRC, he has moved on to
consultation work on 3D and applied human visual perception. Currently, he is
an adjunct professor at the University of Ottawa.

You might also like