Download as pdf or txt
Download as pdf or txt
You are on page 1of 16

896 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO.

5, SEPTEMBER 2011

A Review of Adaptive Image Representations


Gabriel Peyré

Abstract—Improving the modeling of natural images is impor- Typical inverse problems are super-resolution, inpainting (see
tant to go beyond the state-of-the-art for many image processing for instance [5], [69]) and compressed sensing [11], [34].
tasks such as compression, denoising, inverse problems, and tex- Non-Adaptive Regularization: Classical regularization
ture synthesis. Natural images are composed of intricate patterns
such as regular areas, edges, junctions, oriented oscillations, and theory makes use of a prior that is intended to
textures. Processing efficiently such a wide range of regularities re- be low for the images one is interested in. Priors based on
quires methods that are adaptive to the geometry of the image. This derivatives of the image impose some smoothness on the
adaptivity can be achieved using sparse representations in a redun- recovered image. The simplest prior is the norm of the gra-
dant dictionary. The geometric adaptivity is important to search dient , where is a finite difference
for efficient representations in a structured dictionary. Another
way to capture this geometry is through non-local interactions be- approximation of the gradient. This corresponds to a Sobolev
tween patches in the image. The resulting non-local energies can be prior that favors uniformly smooth images. The total variation
used to perform an adaptive image restoration. This paper reviews prior, introduced for denoising by Rudin, Osher, and Fatemi
these emerging technics and shows the interplay between sparse [89], reads . It allows one to take into
and non-local regularizations. account discontinuous images, and favors piecewise constant
Index Terms—Adaptivity, approximation, best basis, denoising, images with edge discontinuities of small perimeters. Other
dictionary learning, inverse problems, non-local regularization, classes of priors include sparsity in orthogonal or redundant
sparse regularization, texture synthesis, triangulations.
representations, which are discussed in Section II.
A prior is taken into account during the inversion of
Finding efficient representations for natural images is rele- using a variational minimization
vant for many image processing tasks. Sparse and non-local reg-
ularizations have recently emerged as unifying concepts to for- (1)
malize this notion of geometric representations. These methods
have been proved useful to perform compression, denoising, in- where weights the fidelity to the observations versus the
version of linear operators, and texture synthesis. This paper re- regularity of the solution, and should be adapted to the noise
views these technics and shows that the concepts of sparsity and level.
variational minimization are at the heart of the success of these Adaptive Regularization: Instead of using a fixed prior ,
methods. it is possible to enhance the regularization by using a family
of priors . The parameter should be adapted to match
the geometric features of the images of interest. In the practical
I. INTRODUCTION
situations considered in this paper, the parameter constrains
the local position and direction of edges and textures. We will

A
of
LARGE class of image processing problems can be for-
malized as recovering a high-resolution image
pixels from noisy observations , where .
thus loosely refer to this parameter as the “geometry” of the
image.
This parameter takes different forms across the sections of
The forward model assumes that is a degraded version of an this review paper. For instance, it is a quadtree in Section III,
unknown high-resolution image , where . The an association field in Section IV-B, a triangulation in
linear operator models the acquisition process. Section IV-C, a manifold in Section V-A, and a weighted
If , the recovery of a good approximation of cor- graph in Section V. This parameter always caries some geomet-
responds to removing noise. When is ill-posed, the inverse rical information about the location of edges or the direction of
problem of recovering a high-resolution image is more difficult. textures.
The non-adaptive regularization (1) is turned into an adaptive
recovery by minimizing
Manuscript received October 05, 2010; revised December 31, 2010; ac-
cepted February 21, 2011. Date of publication February 28, 2011; date of where
current version August 17, 2011. This work was supported by the French
“Agence Nationale de la Recherche” (ANR), under Grant NatImages
(ANR-08-EMER-009), “Adaptivity for natural images and texture represen- (2)
tations”. The associate editor coordinating the review of this manuscript and
approving it for publication was Dr. Jean-Luc Starck.
The author is with the CEREMADE CNRS-Université Paris-Dauphine, The additional term allows one to introduce constraints on
75775 Paris, Cedex 16, France (e-mail: gabriel.peyre@ceremade.dauphine.fr).
the set of admissible geometries.
Color versions of one or more of the figures in this paper are available online
at http://ieeexplore.ieee.org. For a fixed , the energy is usually convex, but is
Digital Object Identifier 10.1109/JSTSP.2011.2120592 non-convex jointly on . One thus has to resort to iterative

1932-4553/$26.00 © 2011 IEEE


PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 897

schemes that progressively estimate both the image and the ge-
ometry, and converge to a stationary point of .
The following sections describe several instances of this
adaptive regularization framework (2), each time using a dif-
ferent parameterization of the prior adapted to several kinds
of edge or texture features.

II. SPARSE AND ADAPTIVE REPRESENTATIONS


Computing a sparse decomposition corresponds to approxi-
mating an image of pixels as
using a small number of
atoms chosen in a dictionary of atoms
.

A. Nonlinear Approximation
Best Approximation and Thresholding: A sparse approxima-
tion is efficient if the error is small. The best ap-
proximation with nonzero coefficients in the dictionary
can be found by solving a variational minimiza-
tion where Fig. 1. Examples of images within several complexity classes.

them to capture the directionality of regular edges. The curvelet


(3) frame of Candès and Donoho [10] enjoys an error decay of
for images that are outside edges, thus
In the case where is an orthogonal basis enhancing over the wavelet approximation. A discrete imple-
of atoms, the problem (3) is re-written as mentation of the curvelet transform corresponds to the projec-
tion of the image on a tight frame [12]. This curvelets
approximation has also been used for inverse problems regular-
ization [13]. For smoothly varying textures, a fixed waveatom
where (4) frame [29] provides a small approximation error.
To approximate efficiently images with complicated edges
for a value of adapted so that . This is the vari- and textures, more redundant dictionaries are needed. To keep
ational regularization (1) for the identity operator , no the computation of the approximation tractable, it is how-
noise , , and the prior . Although this prior ever important to use a dictionary with some structure. Many
is non-convex, a global minimizer is found using a hard classical image processing dictionaries (such as Fourier and
thresholding that selects the inner products with largest ampli- wavelets) are structured because the atoms are parameterized
tudes, see for instance [65] by their frequency, position, and size. Imposing similar struc-
tures in redundant dictionaries is challenging but is important
to obtain fast and efficient algorithms.

and (5) B. Sparse Approximation Over Redundant Dictionaries


For an arbitrary redundant dictionary, computing the best
Computing sparse approximations in orthogonal bases -term approximation (3) is however NP-hard [27], and one
is at the heart of many efficient compression algorithms. For thus needs to resort to greedy schemes [68] or minimization.
instance JPEG-2000 uses a nearly orthogonal wavelet basis and Sparse Approximation: A sparse approximation is ob-
is a state-of-the-art coding scheme; see [65]. tained by convexifying the pseudo norm in (3), and solving
Decay of Approximation Error: For several classes of the following basis pursuit denoising convex problem [15],
images, one can prove the asymptotic optimality of such a non- , where
linear approximation scheme (5) in well chosen ortho-bases.
Uniformly images enjoy an optimal asymptotic approx-
(6)
imation error decay in a Fourier
basis. Piecewise regular images with discontinuity curves
of finite perimeters enjoy an optimal asymptotic decay where is adapted so that and where the
in a wavelet basis. Fig. 1 shows norm is defined as . Many algorithms have
some examples of smooth and discontinuous images. been proposed to solve the convex problem (6), for instance it-
Wavelets are however not optimal for piecewise regular im- erative thresholding methods [24], [25]. Sparse coding in re-
ages whose edges are regular. The error for such a cartoon image dundant dictionaries finds application to compression of sounds
decays as . The square support of wavelet atoms forbid [52] and images [46].
898 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

if the sorted magnitudes decay like ,


then the coding-decoding error satisfies

(7)

see [65]. Note that this decay of the coefficients implies


. This shows that a basis that is
efficient to perform nonlinear image approximation is able
to efficiently perform image compression by support coding.
State-of-the-art coders such as JPEG-2000 [65] use advanced
conditional coding schemes that exploit statistical redundan-
cies among the coefficients. This improves the decay of the
distortion with respect to a naive support coder.
This result does not extend to an arbitrary dictionary, since it
might be too costly to code the coefficients. It extends to dic-
tionaries of ortho-bases detailed in Section III, if the number of
atoms within the dictionary is not too large.
Denoising: Denoising corresponds to estimating an image
from a noisy observation , where is a
deterministic, unknown, clean image. A standard assumption is
Fig. 2. Image compression using wavelet support coding. that is a Gaussian white noise of variance .
A popular nonlinear denoising method is the hard thresh-
olding over an orthogonal basis , where
Such sparse approximation can be shown to be close to op-
is defined in (5). This corresponds to the minimization of (1)
timal for a restricted class of dictionaries, see for instance [98].
for and the prior. It was introduced by Donoho
Most image processing dictionaries are however too redundant
and Johnstone [36], and uses a well chosen proportional
to fit within this theoretical analysis. Furthermore, for highly re-
to the noise level . In an orthogonal basis , the decay
dundant dictionaries, solving (6) is computationally demanding.
implies
Redundant and Structured Dictionaries: To avoid these pit-
falls, one needs to use structured dictionaries which enable a
theoretical analysis of the approximation performance for re- (8)
stricted class of images, and also enable a fast search for nearly
optimal approximations. Sections III and IV detail several struc- if one uses the universal threshold ; see [36].
tured dictionaries, together with adaptive strategies to perform Here denotes the expectation with respect to the noise
this search. random variable , and is the best terms approximation
of in . This shows that the average denoising error decays
fast to zero with if is able to efficiently approximate .
Denoising by thresholding is extended to redundant frames,
C. Sparse Approximation for Processing
in which case is reconstructed using the dual frame
Building ortho-bases or frames for which the approximation
error decays fast is important for many applications. One can
show that for several image processing problems, the efficiency
of the processing is related to the nonlinear approximation error.
This section only sketches the main results for compression, de- which is the dictionary defined by , where
noising, and inverse problem regularization. It gives informal . Denoising artifacts are reduced by using a
statements that require additional technical hypotheses to be translation invariant redundant frame [23]. Denoising of images
mathematically rigorous. is also enhanced by using statistical models that take into ac-
Compression: Efficient image compression schemes per- count the non-independence of coefficients of natural images;
form a binary coding of quantized coefficients , where see for instance [88].
is an orthogonal basis. An image is coded Inverse Problems: For inverse problems, one measures
with bits, and the decoder computes an approximation of , where is an additive noise, and where
. is ill-conditioned. To recover an approximation of
To exploit the sparsity of the decomposition of the image, from , one replaces the sparse approximation by a regular-
a support coding scheme codes independently the locations of ization , where
the nonzero coefficients and their amplitudes. Fig. 2 shows an
example of compression using this simple coding scheme.
Under some conditions on the basis, which are satisfied by (9)
many classical ortho-bases such as wavelets, one can shows that
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 899

Each sub-space has dimension and is equipped


with one or several orthogonal bases indexed by a token

Having multiple bases at each node increases the expressive-


ness of the dictionary, and is required to perform directional ap-
proximation of cartoon images, that are regular outside a set of
regular edge curves.
Fig. 3. Left: example of dyadic subdivision of in squares ; right:
corresponding quad-tree . A basis is obtained by aggregating bases for that
are leaves of , and for a specific choice of token
at each node of the tree
This is equivalent to computing a sparse decomposition of in
the redundant dictionary of using basis
pursuit denoising (6).
In the special case, where is an orthogonal basis, the
minimization (9) can be rewritten as the variational problem (1) When one does not care about the location of the basis elements
using an prior . in the tree, the basis is written as , where the index
A few theoretical results exist, but they are restricted is with and .
to the case where is a well-conditioned frame [98], This dictionary defines a highly redundant set of atoms
or for compressed sensing [11], [34] when is random. . One could use directly to compute
These results roughly show that in these restricted cases, a sparse approximation using for instance minimization (6).
, where is the variance of the noise, if This is however numerically expensive for large images. Re-
decays fast and if is well chosen. This means that stricting the sparsity to orthogonal bases leads to fast algorithms
the noise is kept under control during the inversion. for structured dictionaries.

B. Best Basis Selection


III. TREE-STRUCTURED REPRESENTATIONS
Generalizing (4) to an adaptive setting, the approximation in
To improve the representation of geometrical images, one can a best basis is defined as being a solution of
use a family of bases instead of a fixed basis. A dictionary of
orthogonal bases is a set of orthogonal bases
of , where is the number of pixels in
the image. Instead of using an a priori fixed basis such as the where (10)
wavelet basis, one chooses a parameter adapted to the
structures of the image to process and then uses the optimized Note that this minimization fits into the general framework (2)
basis . with , , , and . Note also that
We insist on the fact that the parameter has a different is the best terms approximation of with any basis in ,
meaning across the sections of this paper. In this section, is where is the number of nonzero coefficients of in .
a quadtree, but this will not be the case anymore in the other According to (5), for each , the optimal image that mini-
sections. mizes is . The minimization (10) can thus
be simplified as , where

A. Quadtree-Based Dictionaries (11)


To enable the fast optimization of a parameter adapted to
a given image to process, each is constrained to be where . This kind of Lagrangian is effi-
a quadtree. The quadtree that parametrizes a basis defines ciently optimized using a dynamic search algorithm, originally
a dyadic segmentation , where each presented by Coifman et al. [20]. It is a particular instance of the
is a square of size . Fig. 3 shows an example of Classification and Regression Tree (CART) algorithm of Brei-
such a dyadic subdivision. This segmentation is used to partition dman et al. [7] as explained by Donoho [35].
either the spacial domain (local cosine packets and bandlets) or The complexity of the algorithm is proportional to the com-
the frequency domain (wavelet packets). plexity of computing the whole set of inner products
The full tree indexes sub-spaces of for scales in the dictionary. For several dictionaries, such as
and position , that obey a refinement rela- those considered in this section, a fast algorithm performs this
tionship , the sum being orthogonal. computation in operations, where is the total number of
To enrich the representation parameterized by a quadtree, a atoms in . For tree structured dictionaries, this complexity is
token is attached to each leave of the tree. thus , where is the number of tokens asso-
We call such a tree with its token a tagged quadtree, and use the ciated to each leaf of the tree. This is much smaller than the total
notation to refer to both the quadtree and its tokens. number of bases in , which grows exponentially with .
900 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

C. Spacial and Frequency Adaptive Dictionaries


Local cosine and wavelet packets bases are parameterized
by a quadtree tree that segments respectively the image
plane and the frequency plane [20], [22], [65].
The wavelet packet construction is generalized by consid-
ering non stationary (NS) wavelet packets [18], that apply dif-
ferent quadrature mirror filters at each scale of the tree. An NS
wavelet packet basis is parameterized by a tagged quadtree
, where each node is tagged by a filter. The selection of a best Fig. 4. Example of subdivision adapted to a cartoon image . The basis
NS wavelet packet basis is obtained by a dynamic program- minimizes .
ming algorithm detailed in [76] that generalizes the best basis
algorithm. scale of the wavelet transform. The orthogonal bases are
computed using discrete oriented orthogonal bases. If is a
D. Geometry Adaptive Dictionaries function that is outside a set of -regular edge curves, the
approximation in the best bandlet basis minimizing (11)
Geometric images with regular edges are difficult to approx-
satisfies , where is the number of
imate using a wavelet basis. Wedglets approximations and ban-
bandlets composing . Such an orthogonal
dlets bases compute a quadtree segmentation of the image
bandlet transform finds applications to image [66] and surface
together with a local estimate of the orientation of the edges that
[84] compression and image denoising [85].
is encoded in the tokens .
Adaptive Wedgelets Approximation: A geometric approxi-
E. Adaptive Tree-Structured Processing
mation is obtained by considering for each node a collec-
tion of discontinuous approximation spaces for the square The processing scenarios described in Section II-C extend
of the segmentation. The wedgelets [33] uses a piecewise con- to the best basis setting. For compression and denoising, the
stant approximation on each side of a straight edge that follows best basis parameter is computed from the (possibly noisy)
the direction indexed by . observed image, as a pre-processing. For inverse problems such
The best wedgelets algorithm optimizes the Lagrangian (11) as inpainting or compressed sensing, where the data might be
over all possible wedgelets segmentation , where the severely degraded, it is not possible to estimate directly from
thresholding approximation is replaced by a piecewise constant the observations. One thus has to use an iterative algorithm that
approximation. This scheme is efficient to approximate a piece- progressively refines the best basis parameter as the quality of
wise constant image whose edges are curves. For such a the reconstructed image increases.
cartoon image, the approximation error with wedgelets de- Adaptive Tree-Structured Compression: For compression ap-
cays like , see [48]. It is also possible plications, one computes the best basis adapted to the image
to consider approximation spaces with higher order polynomials to compress by minimizing the corresponding Lagrangian
to capture arbitrary cartoon images [94]. The computation of the (11), and then binary codes the quantized coefficients
low-dimensional projection can be significantly accelerated; see together with the parameter of the best basis, see [65].
[47]. The coding error decay (7) extends to a dictionary of or-
Adaptive Spatial Bandlet Approximation: The bandlet bases thognonal basis if for some ; see
dictionary was introduced by Le Pennec and Mallat [58], [59]. [65]. The total number of bits is reduced by splitting it as
Bandlets perform an efficient adaptive approximation of images , where is the number of bits to code the
with geometric singularities. This transform has been refined by nonzero coefficients among coefficients, and is the
Mallat and Peyré [66], [84] to obtain a dictionary of regular and number of bits to code the quadtree .
orthogonal basis functions. For compressing cartoon images, the best wedgelet
A directional orthogonal basis of elongated directional sub-space provides a distortion rate
wavelets is defined over each square . Keeping only a few . The best bandlet basis further improve
bandlet coefficients and setting the others to zero performs an this rate to for cartoon im-
approximation of the original image that follows the local di- ages that are outside edges.
rection indexed by . The optimal segmentation and tokens Adaptive Tree-Structured Denoising: To perform denoising,
are those that minimize the Lagrangian (11). Fig. 4 shows an the best basis is computed from the noisy observations
example of such a segmentation adapted to a geometric image. by minimizing an empirical Lagrangian ;
Adaptive Approximation Over the Wavelet Domain: Ap- see [57] and [65]. The denoised image is defined by thresholding
plying such an adaptive geometric approximation directly on the noisy observations . The asymptotic op-
the image leads to unpleasant visual artifacts. To overcome timality result (8) extends to a dictionary of orthogonal bases
this issue, one applies a tree structured approximation on the containing atoms by selecting [65].
discrete set of wavelet coefficients. The wedgeprints of Wakin best sub-space was originally designed to perform denoising.
et al. [99] use a vector quantization to extend the wedgelet For piecewise constant function with edges, Wedgelets
scheme to the wavelet domain. The orthogonal bandlets of leads to an optimal decay of the estimation error of
Peyré and Mallat [66] use an adaptive bandlet basis for each , up to logarithmic factor, where is the
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 901

Fig. 5. (a) ( ) original image. (b) ( ) compressed sensing reconstruction using a translation invariant wavelet frame ( dB and 22.1 dB). (c) ( )
reconstruction using iteration in a best bandlet basis ( 39.3 dB and 23.9 dB).

variance of the noise [31], [33]. The best bandlet basis thresh- using a dictionary of bandlet basis. The operator projects an
olding extends this result to a decay of image of pixels on random vectors.
for cartoon images; see [85]. Note that for some specific inverse problems, such as super-
Adaptive Tree-Structured Regularization of Inverse Prob- resoluton, it is possible to estimate directly the local direction of
lems: The resolution (9) of the inverse problem bandlets-like atoms directly from without resorting to a global
is extended to a best basis recovery by solving over both the optimization problem such as (12). This allows to regularize the
image and the quadtree operator inversion with a fast algorithm that produces state-of-
the-art super-resolution results; see [67].

IV. SEMI-STRUCTURED REPRESENTATIONS


where (12) A. Lifting-Based Dictionaries
To enhance the wavelet representation, it is possible to adapt
which corresponds to the general framework introduced in (2) the wavelet filters to the image content, using spatially adaptive
with . filters to avoid creating large wavelet coefficients near edges.
Peyré proposes in [80] an iterative algorithm to compute a The lifting scheme, introduced by Sweldens [96], is an unifying
stationary point of . Starting from some initial , at a step framework to design adaptive biorthogonal wavelets, through
, the estimate is modified as the use of spatially varying local interpolations.
Lifting Scheme: The lifting scheme has been introduced by
Sweldens [96] as a way to decompose wavelet filtering into el-
ementary lifting steps. This allows one to define wavelets on
The best basis parameter is then updated non-translation invariant grids using spatially varying filters. It
can thus be used to define wavelets on surfaces [61], [91] and
on complicated planar triangulations; see Section IV-C.
The initial sampling grid for some is recursively
if , split, for each as a disjoint union ,
where
otherwise. where is a coarser grid, and is a detail grid intended to
store wavelet coefficients. On an image, one can use a traditional
Note that this is a Lagrangian optimization similar to (11), which dyadic refinement so that , or to build non-
can be solved efficiently using the same dynamic programming tensor product wavelets, one can use quincunx sub-sampling
algorithm [20]. The next iterate is defined using a soft thresh- [45], where , but this splitting scheme is general
olding and holds for instance in the case of triangulation refinement;
see Section IV-C.
A biorthogonal wavelet basis is implic-
where itly defined through lifting by computing the wavelet coeffi-
cients for . though the application
If , one can show that is decaying of elementary lifting steps. The lifting algorithm also maintains
through the iterations, and thus converging; see [80]. temporary variables, which are inner products with scaling func-
Fig. 5 shows an example of application of this algorithm to tions for . At the finest scale ,
the recovery from compressed sensing measurements [11], [34] the scaling coefficients are initialized using the data .
902 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

Then, for each , it successively applies a local predictor


to compute the wavelet coefficients, for all

(13)

and an update operator to maintain stability of the transform, for


all

(14)

The association field connects each point of the


detail grid to a set of points on the coarse grid. This
association field is the only degree of freedom of the scheme,
since the weights are usually computed as a polyno-
mial interpolation on the set of points selected by the field, while
is obtained by solving a linear system to impose that
has the same mean and higher order moments as . The
lifting wavelet transform is easily inverted by successively in-
verting steps (14) and (13).
One can chain more than one predict and one update oper- Fig. 6. Examples of grouplet vectors for several positions and scale .
ators, to construct arbitrary biorthogonal wavelets on uniform
1-D grid [26]. Unfortunately, on non-uniform grids, it is impos-
association fields, where some singular point might be-
sible to control properties such as vanishing moments, orthogo-
long to many different fields for various .
nality, and stability of more than one step of predict and update.
Grouplets Over the Image: The grouplet transform also
Lifting Processing: A nonlinear approximation is obtained
makes use of an association field, but it replaces the lifting
by thresholding the lifted wavelet coefficients before re-
computation of wavelet coefficients by an extended Haar trans-
construction. This can be used to perform compression and de-
form, where points in are processed in sequential order to
noising following the schemes detailed in Section II-C. Note
maintain orthogonality.
however that, for arbitrary association fields , the lifted
A grouplet family is an orthogonal basis, or a redundant
wavelets can be unstable and far from orthogonality. This re-
tight frame adapted to the processing of geometric images or
sults in a loss of performance for compression and denoising
regularly oscillating textures, introduced by Mallat [64]. It has
using thresholding.
been used to perform image denoising and super-resolution [64]
Adaptive Predictions: One can adaptively design the associ-
as well as texture inpainting and synthesis [81]. The geometric
ation fields used during the lifting pyramid. One can compute
parameter is a multiscale association field, where
these associations to reduce the length of the wavelet filter near
each is a vector field that follows the geometry of the image
edges [17]. For compression applications, this estimation is per-
at a scale .
formed from already coded coefficients, so that the fields do
For each scale , the flow links the point to , where
not need to be transmitted to the decoder. Such a filter adaptation
. A modified Haar transform progressively
is however no sufficient to capture fully the geometric informa-
extract grouplets detail coefficients along this flow. This corre-
tion. Such schemes are related to adapted subdivision [70].
sponds to the projection of the image on a orthogonal or redun-
To further reduce the distortion of geometric images, the ori-
dant tight frame of . The grouplets atoms
entation of the association fields can be optimized though
at scale have a width of 1 pixel and follow the geometric flow
the scales. Because of the lack of structure of the set of bases
on a width of pixels. Fig. 6 shows examples of grouplets
, computing the fields that minimizes (10) is intractable.
at several scales.
The lifting operators are thus usually computed using heuristics
The selection of an optimal grouplet flow adapted to
to detect the local orientation of edges. The association fields
an image by minimizing a Lagrangian such as (11) is in-
are obtained by minimizing the amplitude of wavelet coeffi-
tractable, because the grouplet dictionary does not have a tree
cients; see for instance [14] and [50] or using local gradient in-
structure. One thus uses a greedy scheme that performs block
formation [54]. These adaptive lifting schemes can be extended
matching to select the patch around that is the closest
to perform adaptive video transform [92] where the lifting op-
to the patch around . Alternative methods compute by
erates through in time by following the optical flow.
integrating a texture flow along a distance of ; see [81].
The grouplet transform is related to the curved wavelet trans-
B. Grouplets
form [100] and the easy path [86]. A distinctive features of grou-
A difficulty with lifted transforms is that they do not guar- plets is that the fields can converge to a singular point, which
antee the orthogonality of the resulting wavelet frame. The sta- means for . This is useful to represent
bility of the transform thus tends to degrade for highly adaptive turbulent textures with complicated singularities.
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 903

Fig. 8. A few iterations of the grouplet inpainting algorithm.

Fig. 7. Example of texture synthesis by statistical modeling of grouplet coef-


ficients. spline function (hat functions centered on the vertices ),
which has dimension .
The best approximation with vertices of is obtained by
Grouplets Over Wavelets: To process image with both ori- an orthogonal projection
ented texture and edges, this grouplet transform can also be ap-
plied over a wavelet transform, so that the wavelet coefficients
at each wavelet scale are projected on a grouplet frame . This
corresponds to the projection of on a multiscale grouplet
frame composed of atoms with varying widths and lengths.
This grouplet transform has been applied to texture synthesis
[81]; see Fig. 7. The synthesis is obtained by sampling grouplet where (15)
coefficients with the same distribution as the coefficient of an
exemplar, and by randomizing the geometry of the flow .
Since is not orthogonal, computing the coefficients re-
Sparse grouplet regularization can be used to perform geo-
quires the resolution of a sparse linear system.
metric inpainting of large missing regions by minimizing (12)
Smooth Image Approximation With Triangulations: For
over the image while adapting progressively the association
uniformly regular functions, one can build isotropic triangula-
field ; see [81]. Fig. 8 shows an example of iterations of the
tions that generate an error ,
grouplet inpainting algorithm, that performs iteratively a soft
where is the Hessian of the image .
thresholding and an estimation of an adapted grouplet flow .
For a image, one should use anisotropic triangles
whose aspect ratio matches the anisotropy of the image
[4]. A triangle near a point should be aligned
C. Adaptive Triangulations with the principal eigenvector of , and the aspect
ratio should match the ratio of the eigen-
A finite elements approximation first builds a segmentation values of . The approximation error with such
of the image domain into elementary cells, and then perform a adapted anisotropic approximations of a image satis-
low-dimensional polynomial approximation on each cell. The fies ; see for
adaptivity comes from the selection of this segmentation, that instance [74]. This shows that anisotropic approximations
should be adapted to the features of the image. have a better constant than isotropic for the decay of the error
Image Approximation by Triangulations: The parameters .
to adapt a triangulation dictionary is a set of Cartoon Image Approximation With Triangulations: For a
vertices and a set of connections cartoon image, which is a function outside contours, one
that prescribes the connectivity between the points. can also diminish the approximation error using
Triangles are convex hulls of the elements of , and they form anisotropic triangles. Near edges, the anisotropic triangles
a partition of the image domain . should obey the same parabolic scaling as curvelet atoms,
One performs approximation with continuous piecewise . Fig. 9 shows an example of such an adapted
affine function on the triangles, which corresponds to using triangulation. Using such anisotropic triangulations for a car-
an adapted basis of the space of linear toon image leads to an error decay ;
904 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

Fig. 9. Adaptive triangulations for piecewise linear approximations of a


-cartoon image. Fig. 10. Comparison of the adapted triangulation scheme [6] with JPEG-2000,
for the same number of bits, triangles.

see [65]. This improves significantly over the wavelet approxi-


mation decay . from to , and impose that is the Delaunay triangula-
Heuristics and Greedy Algorithms: The issue is that com- tion of . This can be shown experimentally to produce highly
puting the best triangulation that minimizes anisotropic meshes, which can be used to perform compression;
over a discretized domain is NP-hard [1]. There is currently see Demaret et al. [30].
no algorithm that guarantees the error decay Multiscale Approximations on Triangulations: Embedded
for an arbitrary cartoon image. One thus needs to rely triangulations with an increasing number of
on subobtimal greedy schemes. These schemes generate a se- vertices satisfy . This is achieved by imposing that
quence of embedded triangulation by either and that each face of is split into one of more
refinement (increasing values) or coarsening. faces of . The set of triangulation defines a hi-
A greedy refinement scheme starts by a simple fixed triangu- erarchical segmentation of the image that can be encoded using
lation of the squares , and iteratively add one or several a tree, generalizing the quadtree construction of Section III-A.
vertices to to obtain a triangulation that minimizes the A biorthogonal wavelet basis is constructed
error . The faces needs also to be updated using a basis of the detail space that satisfies
to obtain . . The decomposition of an image on a
Delaunay refinement introduced by Ruppert [90] and Chew biorthogonal wavelet basis defined on embedded triangulations
[16], proceeds by inserting a single point, that is imposed to be can be computed using lifting steps; see [61] and [91].
a circumcenter of one triangle, and also impose that is a De- Regular splits of orthogonal triangles lead to isotropic adap-
launay triangulation of . This constraint accelerates the search tive triangulations [32]. Splitting triangles according to a well-
and also leads to triangles with provably good isotropic aspect chosen median leads to anisotropic triangulations that exhibit
ratio, which might be useful to compute the approximation of optimal aspect ratio for smooth images; see [19].
the solution of an elliptic PDE on the mesh grid. For image ap-
proximation, one however needs to design aniosotropic trian-
V. PATCH-BASED REPRESENTATIONS
gulations, which requires to modify the notion of circumcenter
using an anisotropic metric [6], [49]. Fig. 10 shows an example A new class of adaptive image processing have emerged re-
of adaptive triangulation produced by the greedy anisotropic re- cently by considering non-local interactions between patches
finement scheme of [6]. Such a refinement scheme can be used in an image. These interactions are used to perform non-local
to perform image compression, and gives results superior to filtering, sparse coding with dictionary learning, and non-local
JPEG-2000 for cartoon images with regular edges. Other refine- regularization of inverse problems.
ment schemes are possible, such as for instance edge bisection,
for which the optimal error decay is known to hold for smooth A. Manifold Geometry of Patches
convex functions [74].
Triangulation thinning algorithms start with a fine scale trian- Manifold of Patches: A patch of size is extracted
gulation of and progressively remove either a vertex, around a pixel in an image by defining for all
an edge or a face to increase as slowly as possible the approxi-
mation error until vertices remain, see for
instance [37]. One can for instance remove a single vertex to go (16)
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 905

Fig. 12. Example of recovery from compressed sensing measure-


ments, using sparsity in a translation invariant wavelet frame, and using (17) for
the manifold of edges.

Fig. 11. Display of the patch manifold that has the topology of a cylinder for
binary edges.

The set of patches extracted from natural images


often depends on a small number of hidden variables. This is for
instance the case near geometric edges or directional textures.
For some simple classes of geometric images and textures,
this set of patches can be shown to be close to a low-dimen-
sional smooth manifold of dimension . This
is for instance the case for binary cartoon images, where edge Fig. 13. Local Fourier analysis of an image containing edges and textures.
patches can be parameterized by the orientation of the closest
edge, and the distance to this edge. The resulting manifold has Locally parallel textures are oscillating in a direction indi-
the topology of a cylinder, as shown in Fig. 11. Similar mani- cated at a point by a smooth vector field . An adaptive
folds can be use to model more general cartoon images and ori- texture norm makes use of the regularity of the patterns and en-
ented textures; see [78]. hances the separation with respect to non-adaptive norms that
Manifold Regularization: The manifold model assumes that favor arbitrary oscillations. Fig. 13 shows the local Fourier ex-
the patches are close to the manifold , and a variational pansion of an image mixing edges and textures. This local
energy quantifies the deviation from this model analysis is obtained by projecting on a local Fourier frame
. Each atom is localized in a patch around a
where point and has a frequency .
Following Maurel et al. [71], an adaptive texture energy
is defined as a weighted quadratic sum of the local
Fourier coefficients
For simple geometric images and textures, where is explic-
itly known, the energy is a fixed, non-adaptive prior. It can
be used to regularize an inverse problem by min-
imizing
where is a weighing function that is close to zero when
and is rapidly increasing with . This weighting function
forces the local frequencies of the texture to be nonzero around
only for frequencies close to . This prior is well suited
to model locally parallel textures.
(17) The energy captures the oscillating content of the image.
The cartoon content of the image can be captured using the total
A stationary point of this non-convex energy is computed variation , where is a
using an alternative minimization on and on the patch finite difference approximation of the gradient of at . The
of the manifold; see [78]. Fig. 12 shows an cartoon and the texture priors are integrated into a variational
example of compressed sensing reconstruction using a regu- energy minimization to perform a separation from noisy
larization with a manifold of edges. If the manifold is also low-dimensional measurement
optimized in (17), it corresponds to an adaptive regularization
that fits into the framework (2).

B. Geometric Texture Regularization


Separating a texture content from the cartoon content of (18)
a noisy image has been emphasized by Meyer
[72] that introduces several norms for oscillating patterns. Sev- where should be adapted to the noise level and where can
eral numerical algorithms have proposed to perform this sepa- be used to encode constraints on the frequency localization of
ration; see for instance [3]. the texture. This separation problem requires the computation of
906 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

Fig. 15. Some vectors of a dictionary learned from a set of patches


extracted from natural images.

point of this minimization; see for instance [83]. When the dic-
tionary is fixed, one performs a sparse coding to find the
coefficients that solve, for each the basis pursuit denoising (6)
Fig. 14. Variational separation using an energy adapted to locally parallel tex-
tures.
(21)
both the components and the adapted geometric flow . It gen-
eralizes the adaptive regularization (2) to the case where several When the coefficients are fixed, the dictionary is the
priors are used. An iterative minimization algorithm computes solution of
a stationary point of ; see [71]. Fig. 14 shows an example of
decomposition obtained by solving (18), using the iden-
tity operator.

which is the minimization of a quadratic convex functional


C. Dictionary Learning
under quadratic convex constraints and can be solved using a
Dictionary Learning for Patch Approximation: Sparse mod- projected gradient descent [83].
eling of the patches requires the design of a dictionary Fig. 15 shows an example of dictionary learned with this al-
adapted to the local geometry of the images gorithm. The exemplars are patches of 16 16 pixels
to process. The dictionary can be learned from a large set of extracted from a library of natural images, and the redundancy
patches . As first proposed by Olshausen of the dictionary is set to .
and Field [75], is optimized to maximize the sparsity of these Application to Denoising and Inverse Problems: This
exemplars; see also [2] and [43] and references therein. learning can be performed jointly with the resolution of an
The optimal dictionary minimizes the average approximation inverse problem by minimizing
error using the basis pursuit sparse coding (6)
(22)

where is defined in (20). The prior enforces the sparsity in


of each patch extracted at some location as defined
in (16)

(19)

and where the constraint on the dictionary is defined as (23)


if ,
(20) Note that this corresponds to the adaptive regularization frame-
otherwise.
work (2) with the replacement of the geometry by a patch
This problem is convex over the dictionary (under convex con- dictionary .
straints) and on the coefficients, but not jointly convex. One can The resolution of (22) requires the minimization of a non-
use an iterative coordinate descent that converges to a stationary convex energy over both the image to recover and the dictionary
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 907

Fig. 16. Comparison of inpainting using sparsity in a fixed dictionary of Fig. 17. Example of non-local spectral vectors for several value of .
wavelets and local DCT (MCA), and using a wavelet basis and a learned
dictionary (adaptive MCA). [62], where the dictionary is trained by enforcing both sparsity
and discrimination of the classes. See [102] for a review of the
to learn. This minimization can be performed alternatively on , application of dictionary learning to computer vision, including
the coefficients involved in (23), and , which are all its application to face recognition.
convex problems. One can show the convergence of this scheme Structured Dictionary Learning: To obtain faster algorithms
toward a stationary point of the energy; see [83]. and enhance the efficiency of the methods on some classes of
An in-depth review of the applications of dictionary learning images, it is possible to consider structured dictionaries. The
to image processing can be found in [41]. For , the min- simplest structure is a union of ortho-bases ,
imization (22) performs state-of-the-art denoising [40]. When where , as already considered in Section III. It
is a masking operator, the minimization (22) performs in- is possible to extend the optimization problem (19) to include a
painting [63], in which case the dictionary can be learned from block-orthogonality constraint; see for instance [60] and [93]. It
exemplar images, and also updated at each step of the minimiza- is also possible to replace the sparsity energy by a weighted
tion. One can also use several dictionaries and perform a joint norm, which corresponds to a Gaussian mixture prior. The
image separation and regularization. This separation process resulting dictionary can be used to perform image compression
extends the original morphological component analysis to an [28] and provide state-of-the-art result for super-resolution and
adaptive setting [83]. Fig. 16 shows the enhancement of the in- deblurring inverse problems [53]. Let us mention that there are
painting obtained using a learned dictionary instead of a fixed more complicated way to structure a dictionary, for instance by
local cosine dictionary. enforcing tree structures in the sparsity patterns of the coeffi-
Application to Computer Graphics and Vision: Texture syn- cients; see [55].
thesis is performed by computing a random synthesized image
whose patches are sparse in a given dictionary ; see [79]. D. Non-Local Methods
This dictionary is learned from patches extracted from an Non-Local Filtering: Non-local filtering have been intro-
exemplar . The synthesis is obtained by initializing to be duced by Buades et al. [8] as an efficient denoising method.
a Gaussian white noise, and then iterating between the sparse A spatially varying filter parameterized by computes
patch coding (21) in using and averaging of the , where characterizes the
reconstructed patches interaction between the features around point and around
point .
An adaptive non-local filtering computes the weights
from the image to process

where is the number of patches that overlap at a point . where


Dictionary learning over sparse patches also allows to per-
form texture classification and segmentation [79], [95]. A dic-
(24)
tionary is learned for each texture class, and a pixel is assigned
to the class that performs the best sparse approximation around
this pixel. A similar idea is applied to perform image recognition where and controls the shape of the weights.
908 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

Fig. 18. Examples of inpainting with 80% of pixels.

As tends to zero, the filtering becomes related to texture For some inverse problems such as deblurring
synthesis method [38], [101]; see Fig. 7, third image from [56], [73] or inpainting small holes [9], it is possible to estimate
the left. Indeed, given an input image , and starting from the graph from the observations . For generic inverse
a random noise image , applying the recursive scheme problems, the graphs needs to be optimized in parallel to the
for a small produces a random image reconstruction. In this setting, the weights are constrained to
visually similar to . be a probability distribution, so that for each .
Non-Local Spectral Bases: This adaptive filtering is con- Following [44], the spread of the weights is controlled using a
nected to adapted decomposition by introducing an adaptive negative entropy with weight
non-local spectral basis [77], [97]. It is defined as the orthog- .
onal eigenvectors basis of the filtering oper- Recovery from the noisy measurements is
ator obtained by minimizing

(26)

and where is the corresponding eigenvalue, assumed as introduced in [44] for inpainting and in [82] for generic
to be sorted in increasing order. The eigenvectors are a gen- inverse problems. Note that this fits into the general adaptive
eralization of Fourier atoms to a spatially adaptive setting. framework (2).
Fig. 17 shows examples of non-local spectral atoms adapted to The minimization of the non-convex problem (26) is per-
the geometry of a cartoon image. They are oscillating near the formed by iteratively minimizing with respect to and to .
boundary of the shape to allow a better reconstruction of this For a fixed , the optimal weights minimizing (26) are defined
singularity. as in (24). For a fixed , the optimal image is computed by min-
An adaptive denoising of a noisy observations imizing a convex problem. One can shows that for , this
is obtained by computing the non-local graph and then scheme converges to a stationary point of defined in (26); see
thresholding the coefficients in this basis ; see [82].
[77]. Fig. 18 shows a comparison between this approach and total
Non-Local Regularization: The non-local filtering variation and wavelet sparsity for the recovery from missing
is related to the minimization of a non-local energy random pixels.
; see for instance [21], [42], and [51].
To tackle the regularization of an inverse problem , VI. CONCLUSION
this energy is replaced by To conclude this paper, we would like to recall the main
strengths and limitations of the methods detailed in this review.
(25)
— Wedglets (Section III-D) are efficient to approximate car-
toon images, but might lead to visual artifacts because of
see [44] and [82]. For , it is a generalization of the Sobolev the discontinuous segmentation of the image domain.
prior to the non-local setting. For , it extends the total — Bandlets (Section III-D) adopt a strategy similar to
variation. wedgelets, but can be applied over a wavelet domain to
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 909

produce an orthogonal expansion with reduced spacial [2] M. Aharon, M. Elad, and A. M. Bruckstein, “The K-SVD: An algo-
artifacts. This is useful in particular for compression rithm for designing of overcomplete dictionaries for sparse represen-
tation,” IEEE Trans. Signal Process., vol. 54, no. 11, pp. 4311–4322,
applications. Bandlets and wedgelets are however highly Nov. 2006.
constrained constructions, and might not be the best [3] J.-F. Aujol and A. Chambolle, “Dual norms and image decomposition
choice to process complicated natural images. models,” Int. J. Comput. Vis., vol. 63, no. 1, pp. 85–104, Jun 2005.
[4] I. Babuška and A. K. Aziz, “On the angle condition in the finite element
— Lifting (Section IV-A) is a simple way to produce adap- method,” SIAM J. Numer. Anal., vol. 13, no. 2, pp. 214–226, Apr. 1976.
tive directional wavelets. The large number of degrees of [5] M. Bertalmio, G. Sapiro, V. Caselles, and C. Ballester, “Image in-
freedom in these transforms makes the use of heuristic al- painting,” in Proc. Siggraph’00, 2000, pp. 417–424.
[6] S. Bougleux, G. Peyré, and L. Cohen, “Image compression with
gorithms mandatory, and might be an obstacle for their use geodesic anisotropic triangulations,” in Proc. ICCV’09, 2009, pp.
beyond compression applications. 2343–2348.
— Grouplets (Section IV-B) share similarities with lifting [7] L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Stone, Classifica-
tion and Regression Trees. Belmont, CA: Wadsworth, 1984.
transforms, but provide a conservation of energy, which [8] A. Buades, B. Coll, and J. M. Morel, “A review of image denoising
makes them suitable for inverse problems and texture algorithms, with a new one,” SIAM Multiscale Modeling Simul., vol.
synthesis. 4, no. 2, pp. 490–530, 2005.
[9] A. Buades, B. Coll, J.-M. Morel, and C. Sbert, “Self similarity driven
— Triangulations (Section IV-C) are popular for the resolu- demosaicking,” IEEE Trans. Image Process., vol. 18, no. 6, pp.
tion of PDEs using mesh adaptation. Several algorithms 1192–1202, Jun. 2009.
have been proposed to generate anisotropic triangulations, [10] E. Candès and D. Donoho, “New tight frames of curvelets and optimal
representations of objects with piecewise singularities,” Commun.
which are efficient to approximate cartoon images. Trian- Pure Appl. Math., vol. 57, no. 2, pp. 219–266, 2004.
gulations are however mostly restricted to compression ap- [11] E. Candès, J. Romberg, and T. Tao, “Robust uncertainty principles:
plications. Exact signal reconstruction from highly incomplete frequency infor-
mation,” IEEE Trans. Inf. Theory, vol. 52, no. 2, pp. 489–509, Feb.
— Patch manifolds (Section V-A) provide highly constrained 2006.
models that are only suitable for simple images. While [12] E. J. Candès, L. Demanet, D. L. Donoho, and L. Ying, “Fast discrete
manifolds cannot be used for compression they are able curvelet transforms,” SIAM Multiscale Modeling Simul., vol. 3, no. 5,
pp. 861–899, 2003.
to regularize inverse problems. [13] E. J. Candès and D. L. Donoho, “Recovering edges in ill-posed inverse
— Texture norms (Section V-B) offer image models tailored problems: Optimality of curvelet frames,” Ann. Statist., vol. 30, no. 3,
for oscillating and oriented patterns. They are conceptually pp. 784–842, 2000.
[14] V. Chappelier and C. Guillemot, “Oriented wavelet transform for
simpler than grouplets models, but are restricted to locally image compression and denoising,” IEEE Trans. Image Process., vol.
parallel textures. Their use is mostly restricted to texture/ 15, no. 10, pp. 2892–2903, Oct. 2006.
cartoon separation and inpainting, and they should be used [15] S. S. Chen, D. L. Donoho, and M. A. Saunders, “Atomic decomposition
by basis pursuit,” SIAM J. Sci. Comput., vol. 20, no. 1, pp. 33–61, 1999.
in conjunction with an other model to capture the cartoon [16] L. P. Chew, “Guaranteed-quality mesh generation for curved surfaces,”
content. in SCG ’93: Proc. 9th Annu. Symp. Comput. Geom., New York, 1993,
— Dictionary learning (Section V-C) introduces the impor- pp. 274–280, ACM.
[17] R. L. Claypoole, G. M. Davis, W. Sweldens, and R. G. Baraniuk, “Non-
tant concept of training the representation from examples. linear wavelet transforms for image coding via lifting,” IEEE Trans.
This is a promising area of research, with state-of-the-art Image Process., vol. 12, no. 12, pp. 1449–1459, Dec. 2003.
results in denoising and possible extensions to inverse [18] A. Cohen and N. Dyn, “Nonstationary subdivision schemes and
multiresolution analysis,” SIAM J. Math. Anal., vol. 27, no. 6, pp.
problems and texture/cartoon separation. Their use is until 1745–1769, 1996.
now restricted to small patches because of the difficult [19] A. Cohen, N. Dyn, F. Hecht, and J.-M. Mirebeau, “Adaptive multires-
optimization process that is required. olution analysis based on anisotropic triangulations,” Math. Comput.,
2011, preprint: arXiv:1101.1512, to be published.
— Non-local filtering (Section V-D) is easy to use and give [20] R. Coifman and V. Wickerhauser, “Entropy-based algorithms for
denoising results close to the state-of-the-art. It can be ex- best basis selection,” IEEE Trans. Inf. Theory, vol. IT-38, no. 2, pp.
tended to inverse problems regularization which leads to 713–718, Mar. 1992.
[21] R. R. Coifman, S. Lafon, A. B. Lee, M. Maggioni, B. Nadler, F. Warner,
good results on some problems but requires a slow opti- and S. W. Zucker, “Geometric diffusions as a tool for harmonic anal-
mization process over the set of patches. ysis and structure definition of data: Diffusion maps,” Proc. Nat. Acad.
We hope that this review has helped the reader to understand the Sci., vol. 102, pp. 7426–7431, May 2005.
[22] R. R. Coifman, G. Matviyenko, and Y. Meyer, “Modulated Malvar-
connections between these different ways of capturing the ge- Wilson bases,” J. Appl. Comput. Harmon. Anal., vol. 4, no. 1, pp.
ometry of images. The large diversity of approaches comes both 58–61, 1997.
from the richness of the geometrical features in natural images [23] R. R. Coifman and D. L. Donoho, “Translation-invariant de-noising,”
in Wavelets and Statistics, Lecture Notes in Statistics. New York:
and from the variety of applications in image processing, com- Springer-Verlag, 1995, pp. 125–150.
puter vision, and graphics. We believe there are many exciting [24] P. L. Combettes and V. R. Wajs, “Signal recovery by proximal forward-
avenues of future works, starting with the difficult problem of backward splitting,” SIAM Multiscale Model. Simul., vol. 4, no. 4, pp.
1168–1200, 2005.
bridging the gap between geometrical representations and auto- [25] I. Daubechies, M. Defrise, and C. De Mol, “An iterative thresholding
matic learning of sparse models. algorithm for linear inverse problems with a sparsity constraint,”
Commun. Pure Appl. Math., vol. 57, pp. 1413–1541, 2004.
[26] I. Daubechies and W. Sweldens, “Factoring wavelet transforms into
lifting steps,” J. Fourier Anal. Appl., vol. 4, no. 3, pp. 245–267, 1998.
[27] G. Davis, S. Mallat, and M. Avellaneda, “Greedy adaptive approxima-
REFERENCES tion,” J. Construct. Approx., vol. 13, no. 1, pp. 57–98, 1997.
[28] X. Delaunay, M. Chabert, V. Charvillat, and G. Morin, “Satellite image
[1] P. Agarwal and S. Suri, “Surface approximation and geometric parti- compression by post-transforms in the wavelet domain,” Elsevier
tions,” SIAM J. Comput., vol. 27, no. 4, pp. 1016–1035, 1998. Signal Process., vol. 90, no. 2, pp. 599–610, 2010.
910 IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, VOL. 5, NO. 5, SEPTEMBER 2011

[29] L. Demanet and L. Ying, “Wave atoms and sparsity of oscillatory pat- [56] S. Kindermann, S. Osher, and P. W. Jones, “Deblurring and denoising
terns,” J. Appl. Comput. Harmon. Anal., vol. 3, no. 23, pp. 368–387, of images by nonlocal functionals,” SIAM Mult. Model. Simul., vol. 4,
2007. no. 4, pp. 1091–1115, 2005.
[30] L. Demaret, N. Dyn, and A. Iske, “Image compression by linear splines [57] H. Krim, D. Tuckier, S. Mallat, and D. Donoho, “Near-optimal risk
over adaptive triangulations,” Signal Process. J., vol. 86, no. 7, pp. for best basis search,” IEEE Trans. Inf. Theory, vol. 45, no. 7, pp.
1604–1616, 2006. 2225–2238, 1999.
[31] L. Demaret, F. Friedrich, H. Führ, and T. Szygowski, “Multiscale [58] E. Le Pennec and S. Mallat, “Bandelet image approximation and com-
wedgelet denoising algorithms,” in Proc. SPIE, Wavelets XI, 2006, pression,” SIAM Multiscale Model. Simul., vol. 4, no. 3, pp. 992–1039,
vol. 5914, p. X1-12. 2005.
[32] R. Distasi, M. Nappi, and S. Vitulano, “Image compression by [59] E. Le Pennec and S. Mallat, “Sparse geometric image representations
B-tree triangular coding,” IEEE Trans. Commun., vol. 45, no. 9, pp. with bandelets,” IEEE Trans. Image Process., vol. 14, no. 4, pp.
1095–1100, Sep. 1997. 423–438, Apr. 2005.
[33] D. Donoho, “Wedgelets: Nearly minimax estimation of edges,” Ann. [60] S. Lesage, R. Gribonval, F. Bimbot, and L. Benaroya, “Learning unions
Statist., vol. 27, no. 3, pp. 859–897, 1999. of orthonormal bases with thresholded singular value decomposition,”
[34] D. Donoho, “Compressed sensing,” IEEE Trans. Inf. Theory, vol. 52, in Proc. IEEE Int. Conf. Acoust., Speech, Signal Process. (ICASSP),
no. 4, pp. 1289–1306, 2006. 2005, vol. 5, pp. v293–v296.
[35] D. L. Donoho, “Cart and best-ortho-basis: A connection,” Ann. Statist., [61] M. Lounsbery, T. D. DeRose, and J. Warren, “Multiresolution analysis
vol. 25, no. 5, pp. 1870–1911, 1997. for surfaces of arbitrary topological type,” ACM Trans. Graphics, vol.
[36] D. L. Donoho and I. Johnstone, “Minimax estimation by wavelet 16, no. 1, pp. 34–73, 1997.
shrinkage,” Ann. Statist., vol. 26, pp. 879–921, 1998. [62] J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman, “Supervised
[37] N. Dyn, D. Levin, and S. Rippa, “Data dependent triangulations for dictionary learning,” in Proc. Adv. Neural Inf. Process. Syst., 2008, pp.
piecewise linear interpolation,” IMA J. Numer. Anal., vol. 10, no. 1, 1033–1040.
pp. 137–154, Jan. 1990. [63] J. Mairal, M. Elad, and G. Sapiro, “Sparse representation for color
[38] A. A. Efros and T. K. Leung, “Texture synthesis by non-parametric image restoration,” IEEE Trans. Image Process., vol. 17, no. 1, pp.
sampling,” in Proc. ICCV ’99: Proc. Int. Conf. Comput. Vis.-Vol. 2, 53–69, Jan. 2008.
1999, p. 1033, IEEE Comput. Soc.. [64] S. Mallat, “Geometrical grouplets,” J. Appl. Comput. Harmon. Anal.,
[39] A. A. Efros and W. T. Freeman, “Image quilting for texture synthesis vol. 26, no. 2, pp. 161–180, 2009.
and transfer,” in Proc. Siggraph’01, Aug. 2001, pp. 341–346. [65] S. Mallat, A Wavelet Tour of Signal Processing, 3rd ed. Amsterdam,
[40] M. Elad and M. Aharon, “Image denoising via sparse and redun- The Netherlands: Elsevier, 2009.
dant representations over learned dictionaries,” IEEE Trans. Image [66] S. Mallat and G. Peyré, “Orthogonal bandlets bases for geometric
Process., vol. 15, no. 12, pp. 3736–3745, Dec. 2006. image approximation,” Commun. Pure Appl. Math., vol. 61, no. 9, pp.
[41] M. Elad, M. A. T. Figueiredo, and Y. Ma, “On the role of sparse and 1173–1212, 2008.
redundant representations in image processing,” Proc. IEEE, vol. 98, [67] S. Mallat and G. Yu, “Super-resolution with sparse mixing estimators,”
no. 6, pp. 972–982, 2010. IEEE Trans. Image Process., vol. 19, no. 11, pp. 2889–2900, 2010.
[42] A. Elmoataz, O. Lezoray, and S. Bougleux, “Nonlocal discrete regular- [68] S. Mallat and Z. Zhang, “Matching pursuits with time-frequency dictio-
ization on weighted graphs: A framework for image and manifold pro- naries,” IEEE Trans. Signal Process., vol. 41, no. 12, pp. 3397–3415,
cessing,” IEEE Trans. Image Process., vol. 17, no. 7, pp. 1047–1060, Dec. 1993.
Jul. 2008. [69] S. Masnou, “Disocclusion: A variational approach using level lines,”
[43] K. Engan, S. O. Aase, and J. H. Husoy, “Method of optimal directions IEEE Trans. Image Process., vol. 11, no. 2, pp. 68–76, Feb. 2002.
for frame design,” in Proc. ICASSP’99, Washington, DC, 1999, pp. [70] B. Matei and A. Cohen, Nonlinear Subdivison Schemes : Applications
2443–2446, IEEE Comput. Soc.. to Image processing, in Tutorials on Multiresolution in Geometric
[44] G. Facciolo, P. Arias, V. Caselles, and G. Sapiro, “Exemplar-based Modelling, A. Iske, E. Quak, and M. S. Floater, Eds. New York:
interpolation of sparsely sampled images,” IMA Preprint Series # 2264, Springer Verlag, 2002, pp. 93–97.
2009. [71] P. Maurel, J. F. Aujol, and G. Peyré, “Locally parallel texture mod-
[45] J. C. Feauveau, “Analyze multirésolution avec un facteur de résolution eling,” SIAM J. Imaging Sci.., vol. 4, no. 1, pp. 413–447, 2010.
,” J. Traitement du Signal, vol. 7, no. 2, pp. 117–128, 1990. [72] Y. Meyer, Oscillating Patterns in Image Processing and Nonlinear
[46] R. M. Figueras i Ventura, P. Vandergheynst, and P. Frossard, “Low- Evolution Equations. Boston, MA: American Mathematical Society,
rate and flexible image coding with redundant representations,” IEEE 2001.
Trans. Image Process., vol. 15, no. 3, pp. 726–739, Mar. 2006. [73] M. Mignotte, “A non-local regularization strategy for image deconvo-
[47] F. Friedrich, L. Demaret, H. Fuhr, and K. Wicker, “Efficient moment lution,” Pattern Recogn. Lett., vol. 29, no. 16, pp. 2206–2212, Dec.
computation over polygonal domains with an application to rapid 2008.
wedgelet approximation,” SIAM J. Sci. Comput., vol. 29, no. 2, pp. [74] J.-M. Mirebeau and A. Cohen, “Greedy bisection generates optimally
842–863, 2007. adapted triangulations,” Laboratoire Jacques-Louis Lions, 2008, Tech.
[48] H. Fuhr, L. Demaret, and F. Friedrich, “Beyond wavelets: New image Rep..
representation paradigms,” Survey article, in Document and Image [75] B. A. Olshausen and D. J. Field, “Emergence of simple-cell receptive-
Compression, 2006, M. Barni and F. Bartolini. field properties by learning a sparse code for natural images,” Nature,
[49] P. L. George, H. Borouchaki, P. J. Frey, P. Laug, and E. Saltel, “Mesh vol. 381, no. 6583, pp. 607–609, Jun. 1996.
generation and mesh adaptivity: Theory, techniques,” in Encyclopedia [76] N. Ouarti and G. Peyré, “Best basis search in a non-stationary wavelet
of Computational Mechanics, E. Stein, R. de Borst, and T. J. R. Hughes, packets dictionary,” in Proc. ICIP 2009, 2009.
Eds. New York: Wiley, 2004. [77] G. Peyré, “Image processing with non-local spectral bases,” SIAM Mul-
[50] O. N. Gerek and A. E. Cetin, “Adaptive polyphase subband decompo- tiscale Model. Simul., vol. 7, no. 2, pp. 703–730, 2008.
sition structures for image compression,” IEEE Trans. Image Process., [78] G. Peyré, “Manifold models for signals and images,” Comput. Vis.
vol. 9, no. 10, pp. 1649–1660, Oct. 2000. Image Understand., vol. 113, no. 2, pp. 249–260, Feb. 2009.
[51] G. Gilboa, J. Darbon, S. Osher, and T. F. Chan, “Nonlocal convex func- [79] G. Peyré, “Sparse modeling of textures,” J. Math. Imaging Vis., vol.
tionals for image regularization,” UCLA CAM Rep. 06-57, 2006. 34, no. 1, pp. 17–31, 2009.
[52] R. Gribonval and E. Bacry, “Harmonic decomposition of audio signals [80] G. Peyré, “Best basis compressed sensing,” IEEE Trans. Signal
with matching pursuit,” IEEE Trans. Signal Process., vol. 51, no. 1, pp. Process., vol. 28, no. 5, 2010.
101–111, Jan. 2003. [81] G. Peyré, “Texture synthesis with grouplets,” IEEE Trans. Patt. Anal.
[53] G. Yu, G. Sapiro, and S. Mallat, “Solving inverse problems with piece- Mach. Intell., vol. 32, no. 4, pp. 733–746, Apr. 2010.
wise linear estimators: From Gaussian mixture models to structured [82] G. Peyré, S. Bougleux, and L. D. Cohen, “Non-local regularization of
sparsity,” 2011, preprint: arxiv.org/abs/1006.3056, to be published. inverse problems,” Inverse Problems Imag., 2011, to be published.
[54] H. J. Heijmans, B. Pesquet-Popescu, and G. Piella, “Building nonre- [83] G. Peyré, J. Fadili, and J.-L. Starck, “Learning the morphological di-
dundant adaptive wavelets by update lifting,” J. Appl. Comput. versity,” SIAM J. Imag. Sci., vol. 3, no. 3, pp. 646–669, 2010.
Harmon. Anal., vol. 18, no. 3, pp. 252–281, May 2005. [84] G. Peyré and S. Mallat, “Surface compression with geometric ban-
[55] R. Jenatton, J. Mairal, G. Obozinski, and F. Bach, “Proximal methods delets,” ACM Trans. Graphics, (SIGGRAPH’05), vol. 24, no. 3, pp.
for hierarchical sparse coding,” Preprint HAL inria-00516723, 2010. 601–608, Aug. 2005.
PEYRÉ: REVIEW OF ADAPTIVE IMAGE REPRESENTATIONS 911

[85] G. Peyré, E. Le Pennec, C. Dossal, and S. Mallat, “Geometric estima- [97] A. D. Szlam, M. Maggioni, and R. R. Coifman, “Regularization on
tion with orthogonal bandlet bases,” in SPIE Wavelet XII, 2007. graphs with function-adapted diffusion processes,” J. Mach. Learn.
[86] G. Plonka, “The easy path wavelet transform: A new adaptive wavelet Res., vol. 9, pp. 1711–1739, 2008.
transform for sparse representation of two-dimensional data,” SIAM [98] J. A. Tropp, “Just relax: Convex programming methods for identifying
Multiscale Model. Simul., vol. 7, no. 3, pp. 1474–1496, 2009. sparse signals in noise,” IEEE Trans. Inf. Theory, vol. 52, no. 3, pp.
[87] J. Portilla and E. P. Simoncelli, “A parametric texture model based on 1030–1051, Mar. 2006.
joint statistics of complex wavelet coefficients,” Int. J. Comput. Vis., [99] M. Wakin, J. Romberg, H. Choi, and R. Baraniuk, “Wavelet-domain
vol. 40, no. 1, pp. 49–70, 2000. approximation and compression of piecewise smooth images,” IEEE
[88] J. Portilla, V. Strela, M. Wainwright, and E. P. Simoncelli, “Image de- Trans. Image Process., vol. 15, no. 5, pp. 1071–1087, May 2006.
noising using a scale mixture of Gaussians in the wavelet domain,” [100] D. M. Wang, L. Zhang, A. Vincent, and F. Speranza, “Curved wavelet
IEEE Trans. Image Process., vol. 12, no. 11, pp. 1338–1351, Nov. transform for image coding,” IEEE Trans. Image Process., vol. 15, no.
2003. 8, pp. 2413–2421, Aug. 2006.
[89] L. I. Rudin, S. Osher, and E. Fatemi, “Nonlinear total variation based [101] L.-Y. Wei and M. Levoy, “Fast texture synthesis using tree-structured
noise removal algorithms,” Phys. D, vol. 60, no. 1–4, pp. 259–268, vector quantization,” in Proc. 27th Annu. Conf. Comput. Graphics In-
1992. teract. Tech. SIGGRAPH’00: , 2000, pp. 479–488, ACM Press/Ad-
[90] J. Ruppert, “A Delaunay refinement algorithm for quality 2-dimen- dison-Wesley.
sional mesh generation,” J. Algorithms, vol. 18, no. 3, pp. 548–585, [102] J. Wright, Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, and S. Yan,
1995. “Sparse representation for computer vision and pattern recognition,”
[91] P. Schröder and W. Sweldens, “Spherical wavelets: Efficiently repre- Proc. IEEE, vol. 98, no. 6, pp. 1031–1044, Jun. 2010.
senting functions on the sphere,” in Proc. SIGGRAPH 95, 1995, pp.
161–172. Gabriel Peyré graduated from Ecole Normale
[92] A. Secker and D. Taubman, “Lifting-based invertible motion adap- Supérieure de Cachan, Cahan, France, in 2003 and
tive transform (LIMAT) framework for highly scalable video compres- received the Ph.D. degree in applied mathematics
sion,” IEEE Trans. Image Process., vol. 12, no. 12, pp. 1530–1542, from École Polytechnique, Paris, France, in 2005.
Dec. 2003. His research interests include geometrical repre-
[93] O. G. Sezer, O. Harmanci, and O. G. Guleryuz, “Sparse orthonormal sentations for image processing with applications
transforms for image compression,” in Proc. IEEE Int Conf. Image in computer and biological vision. He worked with
Process. (ICIP), 2008, pp. 149–152. S. Mallat on image compression with bandlets
[94] R. Shukla, P. L. Dragotti, M. Do, and M. Vetterli, “Rate distortion op- during the Ph.D. degree. Since 2006, he has been
timized tree structured compression algorithms for piecewise smooth a Researcher at the Centre Nationale de Recherche
images,” IEEE Trans. Image Process., vol. 14, no. 3, pp. 343–359, Mar. Scientifique (CNRS), working in Ceremade, Univer-
2005. sity Paris-Dauphine. Since 2005, he has coauthored 22 papers in international
[95] K. Skretting and J. H. Husoy, “Texture classification using sparse frame journals, 35 conference proceedings in top vision and image processing
based representations,” EURASIP J. Appl. Signal Process., vol. 2006, conferences, and two books. He is the main contributor of the third edition of
no. 1, p. 102, 2006. the book Wavelet Tour Of Signal Processing (Academic, 2008) by S. Mallat.
[96] W. Sweldens, “The lifting scheme: A custom-design construction of He is the creator of the “Numerical tours of signal processing” (www.numer-
biorthogonal wavelets,” J. Appl. Comput. Harmon. Anal., vol. 3, no. 2, ical-tours.com), a popular online repository of Matlab/Scilab resources to teach
pp. 186–200, 1996. modern signal and image processing.

You might also like