Professional Documents
Culture Documents
Plug-and-Play Image Restoration With Deep Denoiser Prior
Plug-and-Play Image Restoration With Deep Denoiser Prior
Abstract—Recent works on plug-and-play image restoration have shown that a denoiser can implicitly serve as the image prior for
model-based methods to solve many inverse problems. Such a property induces considerable advantages for plug-and-play image
restoration (e.g., integrating the flexibility of model-based method and effectiveness of learning-based methods) when the denoiser is
discriminatively learned via deep convolutional neural network (CNN) with large modeling capacity. However, while deeper and larger
CNN models are rapidly gaining popularity, existing plug-and-play image restoration hinders its performance due to the lack of suitable
arXiv:2008.13751v2 [eess.IV] 12 Jul 2021
denoiser prior. In order to push the limits of plug-and-play image restoration, we set up a benchmark deep denoiser prior by training a
highly flexible and effective CNN denoiser. We then plug the deep denoiser prior as a modular part into a half quadratic splitting based
iterative algorithm to solve various image restoration problems. We, meanwhile, provide a thorough analysis of parameter setting,
intermediate results and empirical convergence to better understand the working mechanism. Experimental results on three
representative image restoration tasks, including deblurring, super-resolution and demosaicing, demonstrate that the proposed
plug-and-play image restoration with deep denoiser prior not only significantly outperforms other state-of-the-art model-based methods
but also achieves competitive or even superior performance against state-of-the-art learning-based methods. The source code is
available at https://github.com/cszn/DPIR.
Index Terms—Denoiser Prior, Image Restoration, Convolutional Neural Network, Half Quadratic Splitting, Plug-and-Play
1 I NTRODUCTION
treat the plain learning-based methods as general case of (3). • The HQS-based plug-and-play IR is thoroughly ana-
It is easy to note that one main difference between lyzed with respect to parameter setting, intermediate
model-based methods and learning-based methods is that, results and empirical convergence, providing a better
the former are flexible to handle various IR tasks by simply understanding of the working mechanism.
specifying T and can directly optimize on the degraded • Extensive experimental results on deblurring, super-
image y, whereas the later require cumbersome training resolution and demosaicing have demonstrated the
to learn the model before testing and are usually restricted superiority of the proposed plug-and-play IR with
by specialized tasks. Nevertheless, learning-based methods deep denoiser prior.
can not only enjoy a fast testing speed but also tend to
deliver better performance due to the end-to-end training. In 2 R ELATED W ORKS
contrast, model-based methods are usually time-consuming
Plug-and-play IR generally involves two steps. The first step
with sophisticated priors for the purpose of good perfor-
is to decouple the data term and prior term of the objective
mance [10]. As a result, these two categories of methods
function via a certain variable splitting algorithm, resulting
have their respective merits and drawbacks, and thus it
in an iterative scheme consisting of alternately solving a
would be attractive to investigate their integration which
data subproblem and a prior subproblem. The second step is
leverages their respective merits. Such an integration has
to solve the prior subproblem with any off-the-shelf denois-
resulted in the deep plug-and-play IR method which re-
ers, such as K-SVD [21], non-local means [22], BM3D [23].
places the denoising subproblem of model-based optimiza-
As a result, unlike traditional model-based methods which
tion with learning-based CNN denoiser prior.
needs to specify the explicit and hand-crafted image priors,
The main idea of deep plug-and-play IR is that, with
plug-and-play IR can implicitly define the prior via the de-
the aid of variable splitting algorithms, such as alternating
noiser. Such an advantage offers the possibility of leveraging
direction method of multipliers (ADMM) [11] and half-
very deep CNN denoiser to improve effectiveness.
quadratic splitting (HQS) [12], it is possible to deal with the
data term and prior term separately [13], and particularly,
the prior term only corresponds to a denoising subprob- 2.1 Plug-and-Play IR with Non-CNN Denoiser
lem [14], [15], [16] which can be solved via deep CNN The plug-and-play IR can be traced back to [4], [14], [16].
denoiser. Although several deep plug-and-play IR works In [24], Danielyan et al. used Nash equilibrium to derive an
have been proposed, they typically suffer from the following iterative decoupled deblurring BM3D (IDDBM3D) method
drawbacks. First, they either adopt different denoisers to for image debluring. In [25], a similar method equipped
cover a wide range of noise levels or use a single denoiser with CBM3D denoiser prior was proposed for single im-
trained on a certain noise level, which are not suitable age super-resolution (SISR). By iteratively updating a back-
to solve the denoising subproblem. For example, the IR- projection step and a CBM3D denoising step, the method
CNN [17] denoisers involve 25 separate 7-layer denoisers, has an encouraging performance for its PSNR improve-
each of which is trained on an interval noise level of 2. ment over SRCNN [26]. In [14], the augmented Lagrangian
Second, their deep denoisers are not powerful enough, method was adopted to fuse the BM3D denoiser to solve
and thus, the performance limit of deep plug-and-play IR image deblurring task. With a similar iterative scheme as
is unclear. Third, a deep empirical understanding of their in [24], the first work that treats the denoiser as “plug-and-
working mechanism is lacking. play prior” was proposed in [16]. Prior to that, a similar
This paper is an extension of our previous work [17] plug-and-play idea is mentioned in [4] where HQS algo-
with a more flexible and powerful deep CNN denoiser rithm is adopted for image denoising, deblurring and in-
which aims to push the limits of deep plug-and-play IR painting. In [15], Heide et al. used an alternative to ADMM
by conducting extensive experiments on different IR tasks. and HQS, i.e., the primal-dual algorithm [27], to decouple
Specifically, inspired by FFDNet [18], the proposed deep the data term and prior term. In [28], Teodoro et al. plugged
denoiser can handle a wide range of noise levels via a single class-specific Gaussian mixture model (GMM) denoiser [4]
model by taking the noise level map as input. Moreover, into ADMM to solve image deblurring and compressive
its effectiveness is enhanced by taking advantages of both imaging. In [29], Metzler et al. developed a denoising-based
ResNet [19] and U-Net [20]. The deep denoiser is further approximate message passing (AMP) method to integrate
incorporated into HQS-based plug-and-play IR to show denoisers, such as BLS-GSM [30] and BM3D, for compressed
the merits of using powerful deep denoiser. Meanwhile, sensing reconstruction. In [31], Chan et al. proposed plug-
a novel periodical geometric self-ensemble is proposed to and-play ADMM algorithm with BM3D denoiser for sin-
potentially improve the performance without introducing gle image super-resolution and quantized Poisson image
extra computational burden, and a thorough analysis of recovery for single-photon imaging. In [32], Kamilov et al.
parameter setting, intermediate results and empirical con- proposed fast iterative shrinkage thresholding algorithm
vergence are provided to better understand the working (FISTA) with BM3D and WNNM [10] denoisers for non-
mechanism of the proposed deep plug-and-play IR. linear inverse scattering. In [33], Sun et al. proposed FISTA
The contribution of this work is summarized as follows: by plugging TV and BM3D denoiser prior for Fourier pty-
chographic microscopy. In [34], Yair and Michaeli proposed
• A flexible and powerful deep CNN denoiser is to use WNNM denoiser as the plug-and-play prior for in-
trained. It not only outperforms the state-of-the-art painting and deblurring. In [35], Gavaskar and Chaudhury
deep Gaussian denoising models but also is suitable investigated the convergence of ISTA-based plug-and-play
to solve plug-and-play IR. IR with non-local means denoiser.
3
4 Residual Blocks
4 Residual Blocks
TConv
SConv
Conv
Conv
g Up
lin
sca sca
lin
wn g
Do
Fig. 1. The architecture of the proposed DRUNet denoiser prior. DRUNet takes an additional noise level map as input and combines U-Net [20] and
ResNet [36]. “SConv” and “TConv” represent strided convolution and transposed convolution, respectively.
2.2 Plug-and-Play IR with Deep CNN Denoiser 2.3 Why not Use a Blind Gaussian Denoiser for Plug-
and-Play IR?
With the development of deep learning techniques such as
It is worth to emphasize that the denoiser for plug-and-play
network design and gradient-based optimization algorithm,
IR should be designed for non-blind Gaussian denoising.
CNN-based denoiser has shown promising performance in
The reason is two-fold. First, as will be shown in (6b) of
terms of effectiveness and efficiency. Following its success,
the iterative solution for plug-and-play IR, the sub-problem
a flurry of CNN denoiser based plug-and-play IR works
actually corresponds to a non-blind Gaussian denoising
have been proposed. In [37], Romano et al. proposed explicit
problem with a Gaussian noise level. Second, although
regularization by TNRD denoiser for image deblurring and
the non-blind Gaussian denoising problem could be solved
SISR. In our previous work [17], different CNN denoisers
via the blind Gaussian denoiser, the role of the Gaussian
are trained to plug into HQS algorithm to solve deblurring
denoiser for plug-and-play IR is to smooth out the unknown
and SISR. In [38], Tirer and Giryes proposed iterative de-
noise (e.g., structural noise introduced during iterations)
noising and backward projections with IRCNN denoisers
rather than remove the Gaussian noise. As will be shown
for image inpainting and deblurring. In [39], Gu et al. pro-
in Sec. 6, the noise distribution during iterations is usually
posed to adopt WNNM and IRCNN denoisers for plug-and-
non-Gaussian and varies across different IR tasks and even
play deblurring and SISR. In [40], Tirer and Giryes proposed
different iterations. Moreover, as we will see in Sec. 5.1.4,
use the IRCNN denoisers for plug-and-play SISR. In [41],
while the non-blind Gaussian denoiser can smooth out such
Li and Wu plugged the IRCNN denoisers into the split
non-Gaussian noise by setting a proper noise level, the blind
Bregman iteration algorithm to solve depth image inpaint-
Gaussian denoiser does not have such ability as it can only
ing. In [42], Ryu et al. provided the theoretical convergence
remove the Gaussian-like noise [18].
analysis of plug-and-play IR based on forward-backward
splitting algorithm and ADMM algorithm, and proposed
spectral normalization to train a DnCNN denoiser. In [43], 2.4 Difference to Deep Unfolding IR
Sun et al. proposed a block coordinate regularization-by- It should be noted that, apart from plug-and-play IR, deep
denoising (RED) algorithm by leveraging DnCNN [44] de- unfolding IR [45], [46], [47], [48] can also incorporate the
noiser as the explicit regularizer. advantages of both model-based methods and learning-
Although plug-and-play IR can leverage the powerful based methods. The main difference between them is that
expressiveness of CNN denoiser, existing methods generally the latter interprets a truncated unfolding optimization as
exploit DnCNN or IRCNN which do not make full use of an end-to-end trainable deep network and thus usually
CNN. Typically, the denoiser for plug-and-play IR should produce better results with fewer iterations. However, deep
be non-blind and requires to handle a wide range of noise unfolding IR needs separate training for each task. On the
levels. However, DnCNN needs to separately learn a model contrary, plug-and-play IR is easy to deploy without such
for each noise level. To reduce the number of denoisers, additional training.
some works adopt one denoiser fixed to a small noise
level. However, according to [37] and as will be shown in
Sec. 5.1.3, such a strategy tends to require a large number of 3 L EARNING D EEP CNN D ENOISER P RIOR
iterations for a satisfying performance which would increase Although various CNN-based denoising methods have
the computational burden. While IRCNN denoisers can been recently proposed, most of them are not designed for
handle a wide range of noise levels, it consists of 25 separate plug-and-play IR. In [50], [51], [52], a novel training strategy
7-layer denoisers, among which each denoiser is trained on without ground-truth is proposed. In [53], [54], [55], [56],
an interval noise level of 2. Such a denoiser suffers from two real noise synthesis technique is proposed to handle real
drawbacks. First, it does not have the flexibility to hand a digital photographs. However, from a Bayesian perspective,
specific noise level. Second, it is not effective enough due the denoiser for plug-and-play IR should be a Gaussian
to the shallow layers. Given the above considerations, it is denoiser. Hence, one can add synthetic Gaussian noise to
necessary to devise a flexible and powerful denoiser to boost clean image for supervised training. In [57], [58], [59], [60],
the performance of plug-and-play IR. the non-local module was incorporated into the network
4
TABLE 1
Average PSNR(dB) results of different methods with noise levels 15, 25 and 50 on the widely-used Set12 and BSD68 [3], [44], [49] datasets. The
best and second best results are highlighted in red and blue colors, respectively.
Noise
Datasets BM3D WNNM DnCNN N3 Net NLRN RNAN FOCNet IRCNN FFDNet DRUNet
Level
15 32.37 32.70 32.86 – 33.16 – 33.07 32.77 32.75 33.25
Set12 25 29.97 30.28 30.44 30.55 30.80 – 30.73 30.38 30.43 30.94
50 26.72 27.05 27.18 27.43 27.64 27.70 27.68 27.14 27.32 27.90
15 31.08 31.37 31.73 – 31.88 – 31.83 31.63 31.63 31.91
BSD68 25 28.57 28.83 29.23 29.30 29.41 – 29.38 29.15 29.19 29.48
50 25.60 25.87 26.23 26.39 26.47 26.48 26.50 26.19 26.29 26.59
design for better restoration. However, these methods learn 3.2 Training Details
a separate model for each noise level. Perhaps the most It is well known that CNN benefits from the availability of
suitable denoiser for plug-and-play IR is FFDNet [18] which large-scale training data. To enrich the denoiser prior for
can handle a wide range of noise levels by taking the plug-and-play IR, instead of training on a small dataset that
noise level map as input. Nevertheless, FFDNet only has includes 400 Berkeley segmentation dataset (BSD) images
a comparable performance to DnCNN and IRCNN, thus of size 180×180 [9], we construct a large dataset consisting
lacking effectiveness to boost the performance of plug-and- of 400 BSD images, 4,744 images of Waterloo Exploration
play IR. For this reason, we propose to improve FFDNet Database [65], 900 images from DIV2K dataset [66], and
by taking advantage of the widely-used U-Net [20] and 2,750 images from Flick2K dataset [63]. Because such a
ResNet [19] for architecture design. dataset covers a larger image space, the learned model can
slightly improve the PSNR results on BSD68 dataset [3]
3.1 Denoising Network Architecture while having an obvious PSNR gain on testing datasets from
It is well-known that U-Net [20] is effective and efficient for a different domain.
image-to-image translation, while ResNet [19] is superior As a common setting for Gaussian denoising, the noisy
in increasing the modeling capacity by stacking multiple counterpart y of clean image x is obtained by adding
residual blocks. Following FFDNet [18] that takes the noise AWGN with noise level σ . Correspondingly, the noise level
level map as input, the proposed denoiser, namely DRUNet, map is a uniform map filled with σ and has the same spatial
further integrates residual blocks into U-Net for effective size as noisy image. To handle a wide range of noise levels,
denoiser prior modeling. Note that this work focuses on the noise level σ is randomly chosen from [0, 50] during
providing a flexible and powerful pre-trained denoiser to training. Note that the noisy images are not clipped into the
benefit existing plug-and-play IR methods rather than de- range of [0, 255]. The reason is that the clipping operation
signing new denoising network architecture. Actually, the would change the distribution of the noise, which in turn
similar idea of combining U-Net and ResNet can also be will give rise to inaccurate solution for plug-and-play IR.
found in other works such as [61], [62]. The network parameters are optimized by minimizing the
The architecture of DRUNet is illustrated in Fig. 1. Like L1 loss rather than L2 loss between the denoised image and
FFDNet, DRUNet has the ability to handle various noise its ground-truth with Adam algorithm [67]. Although there
levels via a single model. The backbone of DRUNet is is no direct evidence on which loss would result in better
U-Net which consists of four scales. Each scale has an performance, it is widely acknowledged that L1 loss is more
identity skip connection between 2 × 2 strided convolution robust than L2 loss in handling outliers [68]. Regarding
(SConv) downscaling and 2 × 2 transposed convolution to denoising, outliers may occur during the sampling of
(TConv) upscaling operations. The number of channels in AWGN. In this sense, L1 loss tends to be more stable than
each layer from the first scale to the fourth scale are 64, L2 loss for denoising network training. The learning rate
128, 256 and 512, respectively. Four successive residual starts from 1e-4 and then decreases by half every 100,000
blocks are adopted in the downscaling and upscaling of iterations and finally ends once it is smaller than 5e-7. In
each scale. Inspired by the network architecture design for each iteration during training, 16 patches with patch size
super-resolution in [63], no activation function is followed of 128×128 were randomly sampled from the training data.
by the first and the last convolutional (Conv) layers, as well We separately learn a denoiser model for grayscale image
as SConv and TConv layers. In addition, each residual block and color image. It takes about four days to train the model
only contains one ReLU activation function. with PyTorch and an Nvidia Titan Xp GPU.
It is worth noting that the proposed DRUNet is bias-
free, which means no bias is used in all the Conv, SConv
3.3 Denoising Results
and TConv layers. The reason is two-fold. First, bias-free
network with ReLU activation and identity skip connection 3.3.1 Grayscale Image Denoising
naturally enforces scaling invariance property of many im- For grayscale image denoising, we compared the proposed
age restoration tasks, i.e., f (ax) = af (x) holds true for any DRUNet denoiser with several state-of-the-art denoising
scalar a ≥ 0 (please refer to [64] for more details). Second, methods, including two representative model-based meth-
we have empirically observed that, for the network with ods (i.e., BM3D [23] and WNNM [10]), five CNN-based
bias, the magnitude of bias would be much larger than that methods which separately learn a single model for each
of filters, which in turn may harm the generalizability. noise level (i.e., DnCNN [44], N3 Net [60], NLRN [59],
5
(a) Noisy (14.78dB) (b) BM3D (25.82dB) (c) DnCNN (26.83dB) (d) RNAN (27.18dB) (e) FFDNet (26.92dB) (f) DRUNet (27.31dB)
Fig. 2. Grayscale image denoising results of different methods on image “Monarch” from Set12 dataset with noise level 50.
(a) Noisy (14.99dB) (b) BM3D (28.36dB) (c) DnCNN (28.68dB) (d) FFDNet (28.75dB) (e) IRCNN (28.69dB) (f) DRUNet (29.28dB)
Fig. 3. Color image denoising results of different methods on image “163085” from CBSD68 dataset with noise level 50.
3.3.4 Generalizability to Unseen Noise Level Fig. 4. An example to show the generalizability advantage of the
proposed bias-free DRUNet. The noise level of the noisy image is 200.
In order to show the advantage of the bias-free DRUNet, we
also train a DRUNet+B model whose biases were randomly
initialized from a uniform distribution in [−1, 1]. Fig. 4 biases has a similar performance to bias-free DRUNet for
provides the visual results comparison between different the trained noise levels, however, the bias-free DRUNet can
models on a noisy image with an extremely large unseen enhance the generalizability to unseen noise level. Fourth,
noise level of 200. Note that since DnCNN and IRCNN do the noise level map can improve the performance as it in-
not have the flexibility to change the noise level, we first troduces extra information of the noise. Such a phenomenon
multiply the noisy image by a factor of 0.25 so that the noise has also been reported in [18].
level changes from 200 to 50. We then apply the DnCNN
and IRCNN models for denoising and finally obtain the TABLE 4
denoising results with a multiplication of 4. From Fig. 4, The PSNR results comparison with different cases (i.e, case 1:
we can seen that, even trained on noise level range of DRUNet without residual blocks, case 2: DRUNet with less training data
as in [18], case 3: DRUNet without removing the biases, and case 4:
[0, 50], the bias-free DRUNet can still perform well, whereas DRUNet without taking noise level map as input) with noise levels 15,
DRUNet+B (with biases) introduces noticeable visual arti- 25 and 50 on Set12 dataset.
facts while having a much lower PSNR. As we will see in
Noise
Sec. 3.3.5, DRUNet+B has a comparable performance with Datasets
Level
Case 1 Case 2 Case 3 Case 4 DRUNet
bias-free DRUNet for noise levels in [0, 50]. Thus, we can
15 33.12 33.23 33.24 33.16 33.25
conclude that bias-free DRUNet can enhance the general- Set12 25 30.82 30.92 30.93 30.86 30.94
izability to unseen noise level. Note that whether bias-free 50 27.78 27.87 27.89 27.83 27.90
network can benefit other tasks or not remains further study.
3.3.5 Ablation Study 3.3.6 Runtime, FLOPs and Maximum GPU Memory Con-
In order to further analyze the proposed DRUNet, we have sumption
performed an ablation study to quantify the performance Table 5 reports the runtime, FLOPs and maximum GPU
contribution of different factors such as residual blocks, memory consumption comparison with four representative
training data, biases, and noise level map. Table 4 reports methods (i.e., DnCNN, IRCNN, FFDNet and RNAN) on im-
the comparisons between DRUNet with four different cases, ages of size 256×256 and 512×512 with noise level 50. We do
including case 1: DRUNet without skip connections of the not report other methods such as FOCNet for comparison
residual blocks, case 2: DRUNet with less training data as since they are not implemented in PyTorch for a fair and
in [18], case 3: DRUNet without removing the biases (i.e., easy comparison. Note that, in order to reduce the memory
DRUNet+B), and case 4: DRUNet without taking noise level caused by the non-local module, RNAN splits the input
map as input. By comparison, we can have the following image into overlapped patches with predefined maximum
conclusions. First, the residual blocks can ease the training to spatial size and then aggravates the results to obtain the
get a better performance. Second, the performance on Set12 final denoised image. The default maximum spatial size
dataset tends to be saturated with enough training data as is 10,000 which is equivalent to a size of 100×100. We
the DRUNet model with more training data only improves also compare RNAN∗ which sets maximum spatial size to
the PSNR by an average of 0.01dB. Third, DRUNet with 70,000. As a simple example, RNAN and RNAN∗ splits an
7
1e-2
20
K=8
In order to decouple the data term and prior term of (2), 1e-4
0 4 8 12 16 20 24 28 32 36 40
0
0 4 8 12 16 20 24 28 32 36 40
HQS first introduces an auxiliary variable z, resulting in a # Iterations # Iterations
noise level range of [0, 50] is supposed to be enough for σk . Algorithm 1: Plug-and-play image restoration with
Inspired by such domain knowledge, we can instead set σk deep denoiser prior (DPIR).
and λ to implicitly determine µk . Based on the fact that µk Input : Deep denoiser prior model, degraded
should be monotonically increasing, we uniformly sample image y, degradation operation T , image
σk from a large noise level σ1 to a small one σK in log space. noise level σ , σk of denoiser prior model at
This means that µk can be easily determined via µk = λ/σk2 . k -th iteration for a total of K iterations,
Following [17], σ1 is fixed to 49 while σK is determined by trade-off parameter λ.
the image noise level σ . Since K is user-specified and σK has Output: Restored image zK .
clear physical meanings, they are practically easy to set. As
for the theoretical convergence of plug-and-play IR, please 1 Initialize z0 from y, pre-calculate αk , λσ 2 /σk2 .
refer to [31]. 2 for k = 1, 2, · · · , K do
By far, the remaining parameter for setting is λ. Due to 3 xk = arg minx ky − T (x)k2 + αk kx − zk−1 k2 ; //
the fact that λ comes from the prior term and thus should Solving data subproblem
be fixed, we can choose the optimal λ by a grid search 4 zk = Denoiser(xk , σk ) ; // Denoising with deep
on a validation dataset. Empirically, λ can yield favorable DRUNet denoiser and periodical geometric
performance from the range of [0.19, 0.55]. In this paper, self-ensemble
we fix it to 0.23 unless otherwise specified. It should be 5 end
noted that since λ can be absorbed into σ and plays the
role of controlling the trade-off between data term and prior
term, one can implicitly tune λ by multiplying σ by a scalar. results of xk and zk at intermediate iterations as well as the
To have a clear understanding of the parameter setting, by convergence curves. Note that in order to show the advan-
denoting αk , µk σ 2 = λσ 2 /σk2 and assuming σK = σ = 1, tage of the powerful DRUNet denoiser prior over IRCNN
we plot the values of αk and σk with respect to different denoiser prior, we refer to DPIR with IRCNN denoiser prior
number of iterations K = 8, 24, and 40 in Fig. 5. as IRCNN+.
(a) Blurry image (b) DMPHN (9.74dB) (c) FDN (23.53dB) (d) IRCNN (28.45dB) (e) IRCNN+ (28.14dB) (f) DPIR (30.27dB)
Fig. 7. Visual results comparison of different deblurring methods on Leaves. The blur kernel is visualized in the upper right corner of the blurry
image. The noise level is 7.65(3%).
32
30
28
26
24
22
xk
20 zk
18
16
1 2 3 4 5 6 7 8
(a) x1 (16.34dB) (b) z1 (23.75dB) (c) x4 (22.33dB) (d) z4 (29.37dB) (e) x8 (29.34dB) (f) Convergence curves
Fig. 8. (a)-(e) Visual results and PSNR results of xk and zk at different iterations; (f) Convergence curves of PSNR (y-axis) for xk and zk with
respect to number of iterations (x-axis).
structures and thus can be used to evaluate non-local self- 5.1.2 Intermediate Results and Convergence
similarity prior. For the two blur kernels, they are of size Figs. 8(a)-(e) provide the visual results of xk and zk at
17×17 and 27×27, respectively. As shown in Table 6, we also different iterations on the testing image from Fig. 7, while
consider Gaussian noise with different noise levels 2.55(1%) Fig. 8(f) shows the PSNR convergence curves for xk and
and 7.65(3%). Following the common setting, we synthesize zk . We can have the following observations. First, while
the blurry images by first applying a blur kernel and then (6a) can handle the distortion of blur, it also aggravates
adding AWGN with noise level σ . For the parameters K the strength of noise compared to its input zk−1 . Second,
and σK , they are set to 8 and σ , respectively. For z0 , it is the deep denoiser prior plays the role of removing noise,
initialized as y. leading to a noise-free zk . Third, compared with x1 and
10
x2 , x8 contains more fine details, which means (6a) can where ↓bicubic
s denotes bicubic downsamling with down-
iteratively recover the details. Fourth, according to Fig. 8(f), scaling factor s, it has been revealed that these methods
xk and zk enjoy a fast convergence to the fixed point. would deteriorate seriously if the real degradation model
deviates from the assumed one [79], [80]. To remedy this,
5.1.3 Analysis of the Parameter Setting an alternative way is to adopt a classical but practical
While we fixed the total number of iterations K to be degradation model which assumes the low-resolution (LR)
8 and the noise level in the first iteration σ1 to be 49, image is a blurred, decimated, and noisy version of high-
it is interesting to investigate the performance with other resolution (HR) image. The mathematical formulation of
settings. Table 7 reports the PSNR results with different such degradation model is given by
combinations of K and σ1 on the testing image from Fig. 7.
y = (x ⊗ k) ↓s + n, (11)
One can see that larger σ1 , such as 39 and 49, could result
in better PSNR results. On the other hand, if σ1 is small, where ↓s denotes the standard s-fold downsampler, i.e.,
a large K needs to be specified for a good performance, selecting the upper-left pixel for each distinct s×s patch.
which however would increase the computational burden.
As a result, K and σ1 play an important role for the trade-off In this paper, we consider the above-mentioned two
between efficiency and effectiveness. degradation models for SISR. As for the solution of (6a), the
following iterative back-projection (IBP) solution [25], [81]
TABLE 7 can be adopted for bicubic degradation,
PSNR results with different combinations of K and σ1 on the testing
image from Fig. 7. xk = zk−1 − γ(y − zk−1 ↓bicubic
s ) ↑bicubic
s , (12)
where ↑bicubic
s denotes bicubic interpolation with upscaling
K σ1 = 9 σ1 = 19 σ1 = 29 σ1 = 39 σ1 = 49 factor s, γ is the step size. Note that we only show one
4 20.04 23.27 25.70 27.65 28.96 iteration for simplicity. As an extension, (12) can be further
8 22.50 25.96 28.40 29.89 30.27 modified as follows to handle the classical degradation
24 26.58 29.64 30.06 30.13 30.16 model
40 28.60 29.82 29.92 29.98 30.01
xk = zk−1 − γ (y − (zk−1 ⊗ k) ↓s ) ↑s ⊗ k, (13)
5.1.4 Results of DPIR with Blind DRUNet Denoiser where ↑s denotes upsampling the spatial size by filling the
It is interesting to demonstrate the performance of DPIR new entries with zeros. Especially noteworthy is that there
with blind DRUNet denoiser. For this purpose, we have exists a fast close-form solution to replace the above iterative
trained a blind DRUNet denoiser by removing the noise scheme. According to [82], by assuming the convolution is
level map. With the same parameter setting, we provide the carried out with circular boundary conditions as in deblur-
visual results and PSNR results of xk and zk at different ring, the closed-form solution is given by
iterations of DPIR with the blind DRUNet denoiser on the
!
−1 1 (F(k)d) ⇓s
noisy and blurry Leaves in Fig. 9. Note that the testing image xk = F d − F(k) s ,
is same as in Figs. 7 and 8. By comparison, we can see that αk (F(k)F(k)) ⇓s +αk
DPIR with blind denoiser gives rise to much lower PSNR (14)
values than DPIR with non-blind denoiser. In particular, where d = F(k)F(y ↑s ) + αk F(zk−1 ) and where s de-
some structural artifacts can be observed with a closer look notes distinct block processing operator with element-wise
at the final result z8 . As a result, the non-blind denoiser is multiplication, i.e., applying element-wise multiplication to
more suitable than blind denoiser for plug-and-play IR. the s × s distinct blocks of F(k), ⇓s denotes distinct block
downsampler, i.e., averaging the s × s distinct blocks [83]. It
is easy to verify that (15) is a special case of (14) with s = 1.
It is worth noting that (11) can also be used to solve bicubic
degradation by setting the blur kernel to the approximated
bicubic kernel [83]. In general, the closed-form solution (14)
should be advantageous over iterative solutions (13). The
reason is that the former is an exact solution which contains
one parameter (i.e., αk ) whereas the latter is an inexact
solution which involves two parameters (i.e., number of
(a) x4 (21.66dB) (b) z4 (27.64dB) (c) z8 (25.59dB) inner iterations per outer iteration and step size).
Fig. 9. Visual results and PSNR results of xk and zk at different
For the overall parameter setting, K and σK are set to
iterations of the proposed DPIR with blind DRUNet denoiser. The testing
image is same as in Figs. 7 and 8. 24 and max(σ, s), respectively. For the parameters in (12)
and (13), γ is fixed to 1.75, the the number of inner iterations
required per outer iteration is set to 5. For the initialization
5.2 Single Image Super-Resolution (SISR) of z0 , the bicubic interpolation of the LR image is utilized. In
particular, since the classical degradation model selects the
While existing SISR methods are mainly designed for bicu-
upper-left pixel for each distinct s × s patch, a shift problem
bic degradation model with the following formulation
should be properly addressed. To tackle with this, we adjust
y = x ↓bicubic
s , (10) z0 by using 2D linear grid interpolation.
11
(a) Bicubic (24.82dB) (b) RCAN (24.61dB) (c) MZSR (27.34dB) (d) IRCNN (26.89dB) (e) IRCNN+ (28.65dB) (f) DPIR (29.12dB)
Fig. 11. Visual results comparison of different SISR methods on an image corrupted by classical degradation model. The kernel is shown on the
upper-left corner of the bicubicly interpolated LR image. The scale factor is 2.
30
29
28
27
26
xk
zk
25
24
1 3 6 9 12 15 18 21 24
(a) x1 (24.95dB) (b) z1 (27.24dB) (c) x6 (27.59dB) (d) z6 (28.57dB) (e) x24 (29.12dB) (f) Convergence curves
Fig. 12. (a)-(e) Visual results and PSNR results of xk and zk at different iteration; (f) Convergence curves of PSNR results (y-axis) for xk and zk
with respect to number of iterations (x-axis).
TABLE 8
Average PSNR(dB) results of different methods for single image super-resolution on CBSD68 dataset. The best and second best results are
highlighted in red and blue colors, respectively.
Methods Bicubic RCAN SRFBN MZSR IRCNN IRCNN+DPIR Bicubic RCAN SRFBN IRCNN IRCNN+DPIR Bicubic RCAN SRFBN IRCNN IRCNN+DPIR
Kernel s = 2, σ = 0 s = 3, σ = 0 s = 3, σ = 7.65(3%)
(a) 27.60 29.50 29.39 28.89 29.92 30.00 29.78 25.83 25.02 24.98 26.43 26.56 26.50 24.65 22.77 22.58 25.45 25.58 26.01
(b) 26.14 26.77 26.75 29.45 29.49 30.28 30.16 25.57 27.37 27.31 26.88 27.12 27.08 24.45 24.01 24.04 25.28 25.30 26.10
(c) 25.12 25.32 25.31 28.49 27.75 29.23 29.72 24.92 25.87 25.84 26.56 27.23 27.21 23.94 23.42 23.45 24.61 24.92 25.71
(d) 24.31 24.37 24.35 25.26 26.44 27.82 28.71 24.27 24.69 24.70 25.78 27.08 27.18 23.41 22.76 22.79 23.97 24.63 25.17
(e) 24.29 24.38 24.37 25.48 26.41 27.76 28.40 24.20 24.65 24.64 25.55 26.78 27.05 23.35 22.71 22.73 23.96 24.58 25.09
(f) 24.02 24.10 24.09 25.46 26.05 27.72 28.50 23.98 24.46 24.43 25.44 26.87 27.04 23.16 22.54 22.55 23.75 24.51 25.01
(g) 24.16 24.24 24.22 25.93 26.28 27.86 28.66 24.10 24.63 24.61 25.64 27.00 27.11 23.27 22.64 22.68 23.87 24.59 25.12
(h) 23.61 23.61 23.59 22.27 25.45 26.88 27.57 23.63 23.82 23.82 24.92 26.55 26.94 22.88 22.18 22.20 23.41 24.27 24.60
Bicubic 26.37 31.18 31.07 29.47 30.31 30.34 30.12 25.97 28.08 28.00 27.19 27.24 27.23 24.76 24.21 24.24 24.36 25.61 26.35
TABLE 9
Demosaicing results of different methods on Kodak and McMaster datasets. The best and second best results are highlighted in red and blue
colors, respectively.
Datasets Matlab DDR DeepJoint MMNet RLDD RNAN LSSC IRI FlexISP IRCNN IRCNN+ DPIR
Kodak 35.78 41.11 42.00 40.19 42.49 43.16 41.43 39.23 38.52 40.29 40.80 42.68
McMaster 34.43 37.12 39.14 37.09 39.25 39.70 36.15 36.90 36.87 37.45 37.79 39.39
13
(a) Ground-truth (b) Matlab (33.67dB) (c) DDR (41.94dB) (d) DeepJoint (42.49dB) (e) MMNet (40.62dB) (f) RNAN (43.77dB)
(g) LSSC (42.31dB) (h) IRI (39.49dB) (i) FlexISP (36.95dB) (j) IRCNN (40.18dB) (k) IRCNN+ (40.85dB) (l) DPIR (43.23dB)
Fig. 13. Visual results comparison of different demosaicing methods on image kodim19 from Kodak dataset.
44
42
40
38
36
34
32
xk
30 zk
28
1 4 8 12 16 20 24 28 32 36 40
(a) x1 (33.67dB) (b) z1 (29.21dB) (c) x16 (32.69dB) (d) z16 (31.45dB) (e) x40 (43.18dB) (f) Convergence curves
Fig. 14. (a)-(e) Visual results and PSNR results of xk and zk at different iterations; (f) Convergence curves of PSNR results (y-axis) for xk and zk
with respect to number of iterations (x-axis).
directional difference regression (DDR) [91], deep unfolding others give rise to noticeable artifacts.
majorization-minimization network (MMNet) [92], residual
learning-based joint demosaicing-denoising (RLDD) [93], 5.3.2 Intermediate Results and Convergence
deep joint demosaicing and denoising (DeepJoint) [89], very Figs. 14(a)-(e) show the visual results and PSNR results of xk
deep residual non-local attention network (RNAN) [58] and zk at different iterations. One can see that the DRUNet
learned simultaneous sparse coding (LSSC) [94], iterative denoiser prior plays the role of smoothing out current
residual interpolation (IRI) [95], minimized-Laplacian resid- estimation x. By passing z through (16), the new output
ual interpolation (MLRI) [96], primal-dual algorithm with x obtained by a weighted average of y and z becomes
CBM3D denoiser prior (FlexISP) [15], IRCNN and IRCNN+. unsmooth. In this sense, the denoiser also aims to diffuse y
Note that DDR, MMNet, RLDD, DeepJoint, and RNAN are for a better estimation of missing values. Fig. 14(f) shows the
learning-based methods, while LSSC, IRI, MLRI, FlexISP, PSNR convergence curves of xk and zk . One can see that the
IRCNN, IRCNN+, and DPIR are model-based methods. two PSNR sequences are not monotonic but they eventually
Table 9 reports the average PSNR(dB) results of different converge to the fixed point. Specifically, a decrease of the
methods on Kodak dataset and McMaster dataset. It can be PSNR value for the first four iterations can be observed as
seen that while RNAN and MMNet achieve the best results, the denoiser with a large noise level removes much more
DPIR can have a very similar result and significantly out- useful information than the unwanted artifacts.
performs the other model-based methods. With a stronger
denoiser, DPIR has an average PSNR improvement up to 6 D ISCUSSION
1.8dB over IRCNN+. While the denoiser prior for plug-and-play IR is trained for
Fig. 13 shows the visual results comparison of different Gaussian denoising, it does not necessary mean the noise
methods on a testing image from Kodak dataset. As one of its input (or more precisely, the difference to the ground-
can see, the Matlab’s simple demosaicing method intro- truth) has a Gaussian distribution. In fact, the noise distri-
duces some zipper effects and false color artifacts. Such bution varies across different IR tasks and even different
artifacts are highly reduced by learning-based methods such iterations. Fig. 15 shows the noise histogram of x1 and x8
as DeepJoint, MMNet and RNAN. For the model-based in Fig. 8 for deblurring, x1 and x24 in Fig. 12 for super-
methods, DPIR produces the best visual results whereas the resolution, and x1 and x40 in Fig. 14 for demosaicing. It
14
can be observed that the three IR tasks has very different intermediate results, empirical convergence were provided.
noise distributions. This is intuitively reasonable because The results have demonstrated that plug-and-play image
the noise also correlates with the degradation operation restoration with powerful deep denoiser prior have several
which is different for the three IR tasks. Another interesting advantages. On the one hand, it boosts the effectiveness
observation is that the two noise distributions of x1 and x8 of model-based methods due to the implicit but powerful
in Fig. 15(a) are different and the latter tends to be Gaussian- prior modeling of deep denoiser. On the other hand, with-
like. The underlying reason is that the blurriness caused by out task-specific training, it is more flexible than learning-
blur kernel is gradually alleviated after several iterations. In based methods while having comparable performance. In
other words, x8 suffers much less from the blurriness and summary, this work has highlighted the advantages of deep
thus is dominated by Gaussian-like noise. denoiser based plug-and-play image restoration. It is worth
noting that there also remains room for further study. For ex-
104 104 105
ample, one direction would be how to integrate other types
2 12 6
of deep image prior, such as deep generative prior [97], for
1.5 9 4.5 effective image restoration.
1 6 3
1 6 3
R EFERENCES
0.5 3 1.5
[16] S. V. Venkatakrishnan, C. A. Bouman, and B. Wohlberg, “Plug- [40] T. Tirer and R. Giryes, “Super-resolution via image-adapted de-
and-play priors for model based reconstruction,” in IEEE Global noising cnns: Incorporating external and internal learning,” IEEE
Conference on Signal and Information Processing, 2013, pp. 945–948. Signal Processing Letters, vol. 26, no. 7, pp. 1080–1084, 2019.
[17] K. Zhang, W. Zuo, S. Gu, and L. Zhang, “Learning deep CNN de- [41] Z. Li and J. Wu, “Learning deep cnn denoiser priors for depth
noiser prior for image restoration,” in IEEE Conference on Computer image inpainting,” Applied Sciences, vol. 9, no. 6, p. 1103, 2019.
Vision and Pattern Recognition, July 2017, pp. 3929–3938. [42] E. Ryu, J. Liu, S. Wang, X. Chen, Z. Wang, and W. Yin, “Plug-
[18] K. Zhang, W. Zuo, and L. Zhang, “FFDNet: Toward a fast and and-play methods provably converge with properly trained de-
flexible solution for cnn-based image denoising,” IEEE Transactions noisers,” in International Conference on Machine Learning, 2019, pp.
on Image Processing, vol. 27, no. 9, pp. 4608–4622, 2018. 5546–5557.
[19] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning [43] Y. Sun, J. Liu, and U. Kamilov, “Block coordinate regularization by
for image recognition,” in IEEE Conference on Computer Vision and denoising,” in Advances in Neural Information Processing Systems,
Pattern Recognition, 2016, pp. 770–778. 2019, pp. 380–390.
[20] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional [44] K. Zhang, W. Zuo, Y. Chen, D. Meng, and L. Zhang, “Beyond
networks for biomedical image segmentation,” in International a gaussian denoiser: Residual learning of deep CNN for image
Conference on Medical Image Computing and Computer-Assisted In- denoising,” IEEE Transactions on Image Processing, pp. 3142–3155,
tervention, 2015, pp. 234–241. 2017.
[21] M. Elad and M. Aharon, “Image denoising via sparse and redun- [45] J. Zhang and B. Ghanem, “ISTA-Net: Interpretable optimization-
dant representations over learned dictionaries,” IEEE Transactions inspired deep network for image compressive sensing,” in IEEE
on Image Processing, vol. 15, no. 12, pp. 3736–3745, 2006. Conference on Computer Vision and Pattern Recognition, 2018, pp.
[22] A. Buades, B. Coll, and J.-M. Morel, “A non-local algorithm 1828–1837.
for image denoising,” in IEEE Conference on Computer Vision and [46] H. K. Aggarwal, M. P. Mani, and M. Jacob, “Modl: Model-based
Pattern Recognition, vol. 2, 2005, pp. 60–65. deep learning architecture for inverse problems,” IEEE transactions
[23] K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Image denois- on medical imaging, vol. 38, no. 2, pp. 394–405, 2018.
ing by sparse 3-D transform-domain collaborative filtering,” IEEE [47] W. Dong, P. Wang, W. Yin, G. Shi, F. Wu, and X. Lu, “Denoising
Transactions on Image Processing, vol. 16, no. 8, pp. 2080–2095, 2007. prior driven deep neural network for image restoration,” IEEE
[24] A. Danielyan, V. Katkovnik, and K. Egiazarian, “BM3D frames Transactions on Pattern Analysis and Machine Intelligence, vol. 41,
and variational image deblurring,” IEEE Transactions on Image no. 10, pp. 2305–2318, 2018.
Processing, vol. 21, no. 4, pp. 1715–1728, 2012. [48] C. Bertocchi, E. Chouzenoux, M.-C. Corbineau, J.-C. Pesquet, and
[25] K. Egiazarian and V. Katkovnik, “Single image super-resolution M. Prato, “Deep unfolding of a proximal interior point method for
via BM3D sparse coding,” in European Signal Processing Conference, image restoration,” Inverse Problems, vol. 36, no. 3, p. 034005, 2020.
2015, pp. 2849–2853. [49] D. Martin, C. Fowlkes, D. Tal, and J. Malik, “A database of hu-
[26] C. Dong, C. C. Loy, K. He, and X. Tang, “Image super-resolution man segmented natural images and its application to evaluating
using deep convolutional networks,” IEEE transactions on Pattern segmentation algorithms and measuring ecological statistics,” in
Analysis and Machine Intelligence, vol. 38, no. 2, pp. 295–307, 2016. IEEE International Conference on Computer Vision, vol. 2, July 2001,
[27] A. Chambolle and T. Pock, “A first-order primal-dual algorithm pp. 416–423.
for convex problems with applications to imaging,” Journal of [50] J. Lehtinen, J. Munkberg, J. Hasselgren, S. Laine, T. Karras, M. Ait-
Mathematical Imaging and Vision, vol. 40, no. 1, pp. 120–145, 2011. tala, and T. Aila, “Noise2noise: Learning image restoration with-
[28] A. M. Teodoro, J. M. Bioucas-Dias, and M. A. Figueiredo, “Image out clean data,” in International Conference on Machine Learning,
restoration and reconstruction using variable splitting and class- 2018, pp. 2965–2974.
adapted image priors,” in IEEE International Conference on Image [51] A. Krull, T.-O. Buchholz, and F. Jug, “Noise2void-learning denois-
Processing. IEEE, 2016, pp. 3518–3522. ing from single noisy images,” in IEEE Conference on Computer
[29] C. A. Metzler, A. Maleki, and R. G. Baraniuk, “From denoising Vision and Pattern Recognition, 2019, pp. 2129–2137.
to compressed sensing,” IEEE Transactions on Information Theory, [52] J. Batson and L. Royer, “Noise2self: Blind denoising by self-
vol. 62, no. 9, pp. 5117–5144, 2016. supervision,” in International Conference on Machine Learning, 2019,
[30] J. Portilla, V. Strela, M. J. Wainwright, and E. P. Simoncelli, “Im- pp. 524–533.
age denoising using scale mixtures of gaussians in the wavelet [53] S. Guo, Z. Yan, K. Zhang, W. Zuo, and L. Zhang, “Toward convo-
domain,” IEEE Transactions on Image Processing, vol. 12, no. 11, pp. lutional blind denoising of real photographs,” in IEEE Conference
1338–1351, 2003. on Computer Vision and Pattern Recognition, 2019, pp. 1712–1722.
[31] S. H. Chan, X. Wang, and O. A. Elgendy, “Plug-and-play admm [54] T. Brooks, B. Mildenhall, T. Xue, J. Chen, D. Sharlet, and J. T.
for image restoration: Fixed-point convergence and applications,” Barron, “Unprocessing images for learned raw denoising,” in IEEE
IEEE Transactions on Computational Imaging, vol. 3, no. 1, pp. 84–98, Conference on Computer Vision and Pattern Recognition, 2019, pp.
2016. 11 036–11 045.
[32] U. S. Kamilov, H. Mansour, and B. Wohlberg, “A plug-and-play [55] A. Abdelhamed, M. A. Brubaker, and M. S. Brown, “Noise flow:
priors approach for solving nonlinear imaging inverse problems,” Noise modeling with conditional normalizing flows,” in Proceed-
IEEE Signal Processing Letters, vol. 24, no. 12, pp. 1872–1876, 2017. ings of the IEEE International Conference on Computer Vision, 2019,
[33] Y. Sun, S. Xu, Y. Li, L. Tian, B. Wohlberg, and U. S. Kamilov, pp. 3165–3173.
“Regularized fourier ptychography using an online plug-and-play [56] S. W. Zamir, A. Arora, S. Khan, M. Hayat, F. S. Khan, M.-H.
algorithm,” in IEEE International Conference on Acoustics, Speech and Yang, and L. Shao, “Cycleisp: Real image restoration via improved
Signal Processing, 2019, pp. 7665–7669. data synthesis,” IEEE Conference on Computer Vision and Pattern
[34] N. Yair and T. Michaeli, “Multi-scale weighted nuclear norm Recognition, 2020.
image restoration,” in IEEE Conference on Computer Vision and [57] S. Lefkimmiatis, “Non-local color image denoising with convolu-
Pattern Recognition, 2018, pp. 3165–3174. tional neural networks,” in IEEE Conference on Computer Vision and
[35] R. G. Gavaskar and K. N. Chaudhury, “Plug-and-play ista con- Pattern Recognition, 2017, pp. 3587–3596.
verges with kernel denoisers,” IEEE Signal Processing Letters, [58] Y. Zhang, K. Li, K. Li, B. Zhong, and Y. Fu, “Residual non-
vol. 27, pp. 610–614, 2020. local attention networks for image restoration,” in International
[36] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning Conference on Learning Representations, 2019.
for image recognition,” in IEEE Conference on Computer Vision and [59] D. Liu, B. Wen, Y. Fan, C. C. Loy, and T. S. Huang, “Non-local
Pattern Recognition, 2016, pp. 770–778. recurrent network for image restoration,” in Advances in Neural
[37] Y. Romano, M. Elad, and P. Milanfar, “The little engine that could: Information Processing Systems, 2018, pp. 1673–1682.
Regularization by denoising (red),” SIAM Journal on Imaging Sci- [60] T. Plötz and S. Roth, “Neural nearest neighbors networks,” in
ences, vol. 10, no. 4, pp. 1804–1844, 2017. Advances in Neural Information Processing Systems, 2018, pp. 1087–
[38] T. Tirer and R. Giryes, “Image restoration by iterative denoising 1098.
and backward projections,” IEEE Transactions on Image Processing, [61] Z. Zhang, Q. Liu, and Y. Wang, “Road extraction by deep residual
vol. 28, no. 3, pp. 1220–1234, 2018. u-net,” IEEE Geoscience and Remote Sensing Letters, vol. 15, no. 5,
[39] S. Gu, R. Timofte, and L. Van Gool, “Integrating local and non- pp. 749–753, 2018.
local denoiser priors for image restoration,” in International Con- [62] G. Venkatesh, Y. Naresh, S. Little, and N. E. O’Connor, “A deep
ference on Pattern Recognition, 2018, pp. 2923–2928. residual architecture for skin lesion segmentation,” in Context-
16
Aware Operating Theaters, Computer Assisted Robotic Endoscopy, Clin- [85] Z. Li, J. Yang, Z. Liu, X. Yang, G. Jeon, and W. Wu, “Feedback net-
ical Image-Based Procedures, and Skin Image Analysis. Springer, 2018, work for image super-resolution,” in IEEE Conference on Computer
pp. 277–284. Vision and Pattern Recognition, 2019, pp. 3867–3876.
[63] B. Lim, S. Son, H. Kim, S. Nah, and K. Mu Lee, “Enhanced deep [86] J. W. Soh, S. Cho, and N. I. Cho, “Meta-transfer learning for zero-
residual networks for single image super-resolution,” in IEEE shot super-resolution,” in IEEE Conference on Computer Vision and
Conference on Computer Vision and Pattern Recognition Workshops, Pattern Recognition, 2020, pp. 3516–3525.
2017, pp. 136–144. [87] T. Dai, J. Cai, Y. Zhang, S.-T. Xia, and L. Zhang, “Second-order
[64] S. Mohan, Z. Kadkhodaie, E. P. Simoncelli, and C. Fernandez- attention network for single image super-resolution,” in IEEE
Granda, “Robust and interpretable blind image denoising via bias- Conference on Computer Vision and Pattern Recognition, 2019, pp.
free convolutional neural networks,” in International Conference on 11 065–11 074.
Learning Representations, 2019. [88] H. S. Malvar, L.-w. He, and R. Cutler, “High-quality linear interpo-
[65] K. Ma, Z. Duanmu, Q. Wu, Z. Wang, H. Yong, H. Li, and L. Zhang, lation for demosaicing of bayer-patterned color images,” in IEEE
“Waterloo exploration database: New challenges for image quality International Conference on Acoustics, Speech, and Signal Processing,
assessment models,” IEEE Transactions on Image Processing, vol. 26, vol. 3, 2004, pp. iii–485.
no. 2, pp. 1004–1016, 2017. [89] M. Gharbi, G. Chaurasia, S. Paris, and F. Durand, “Deep joint de-
mosaicking and denoising,” ACM Transactions on Graphics, vol. 35,
[66] E. Agustsson and R. Timofte, “NTIRE 2017 challenge on single
no. 6, pp. 1–12, 2016.
image super-resolution: Dataset and study,” in IEEE Conference on
[90] L. Liu, X. Jia, J. Liu, and Q. Tian, “Joint demosaicing and denoising
Computer Vision and Pattern Recognition Workshops, vol. 3, July 2017,
with self guidance,” in IEEE Conference on Computer Vision and
pp. 126–135.
Pattern Recognition, 2020, pp. 2240–2249.
[67] D. Kingma and J. Ba, “Adam: A method for stochastic optimiza- [91] J. Wu, R. Timofte, and L. Van Gool, “Demosaicing based on
tion,” in International Conference for Learning Representations, 2015. directional difference regression and efficient regression priors,”
[68] C. M. Bishop, Pattern recognition and machine learning. springer, IEEE Transactions on Image Processing, vol. 25, no. 8, pp. 3862–3874,
2006. 2016.
[69] Y. Zhang, K. Li, K. Li, B. Zhong, and Y. Fu, “Residual non- [92] F. Kokkinos and S. Lefkimmiatis, “Iterative joint image demo-
local attention networks for image restoration,” in International saicking and denoising using a residual denoising network,” IEEE
Conference on Learning Representations, 2019. Transactions on Image Processing, vol. 28, no. 8, pp. 4177–4188, 2019.
[70] X. Jia, S. Liu, X. Feng, and L. Zhang, “Focnet: A fractional opti- [93] Y. Guo, Q. Jin, G. Facciolo, T. Zeng, and J.-M. Morel, “Residual
mal control network for image denoising,” in IEEE Conference on learning for effective joint demosaicing-denoising,” arXiv preprint
Computer Vision and Pattern Recognition, 2019, pp. 6054–6063. arXiv:2009.06205, 2020.
[71] R. Franzen, “Kodak lossless true color image suite,” source: [94] J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman, “Non-
http://r0k. us/graphics/kodak, vol. 4, no. 2, 1999. local sparse models for image restoration,” in IEEE International
[72] L. Zhang, X. Wu, A. Buades, and X. Li, “Color demosaicking by lo- Conference on Computer Vision, 2009, pp. 2272–2279.
cal directional interpolation and nonlocal adaptive thresholding,” [95] W. Ye and K.-K. Ma, “Color image demosaicing using itera-
Journal of Electronic Imaging, vol. 20, no. 2, p. 023016, 2011. tive residual interpolation,” IEEE Transactions on Image Processing,
vol. 24, no. 12, pp. 5879–5891, 2015.
[73] C. Dong, Y. Deng, C. C. Loy, and X. Tang, “Compression artifacts
[96] D. Kiku, Y. Monno, M. Tanaka, and M. Okutomi, “Beyond color
reduction by a deep convolutional network,” in IEEE International
difference: Residual interpolation for color image demosaicking,”
Conference on Computer Vision, 2015, pp. 576–584.
IEEE Transactions on Image Processing, vol. 25, no. 3, pp. 1288–1300,
[74] M. Ehrlich, L. Davis, S.-N. Lim, and A. Shrivastava, “Quantization 2016.
guided jpeg artifact correction,” in European Conference on Com- [97] A. Brock, J. Donahue, and K. Simonyan, “Large scale gan training
puter Vision, 2020. for high fidelity natural image synthesis,” in International Confer-
[75] R. Timofte, R. Rothe, and L. Van Gool, “Seven ways to improve ence on Learning Representations, 2018.
example-based single image super resolution,” in IEEE Conference
on Computer Vision and Pattern Recognition, 2016, pp. 1865–1873.
[76] A. Levin, Y. Weiss, F. Durand, and W. T. Freeman, “Understanding
and evaluating blind deconvolution algorithms,” in IEEE Confer-
ence on Computer Vision and Pattern Recognition, 2009, pp. 1964–
1971.
[77] J. Kruse, C. Rother, and U. Schmidt, “Learning to push the limits
of efficient fft-based image deconvolution,” in IEEE International
Conference on Computer Vision, 2017, pp. 4586–4594.
[78] H. Zhang, Y. Dai, H. Li, and P. Koniusz, “Deep stacked hierarchical
multi-patch network for image deblurring,” in IEEE Conference on
Computer Vision and Pattern Recognition, 2019, pp. 5978–5986.
[79] N. Efrat, D. Glasner, A. Apartsin, B. Nadler, and A. Levin,
“Accurate blur models vs. image priors in single image super-
resolution,” in IEEE International Conference on Computer Vision,
2013, pp. 2832–2839.
[80] K. Zhang, X. Zhou, H. Zhang, and W. Zuo, “Revisiting single im-
age super-resolution under internet environment: blur kernels and
reconstruction algorithms,” in Pacific Rim Conference on Multimedia,
2015, pp. 677–687.
[81] M. Irani and S. Peleg, “Motion analysis for image enhancement:
Resolution, occlusion, and transparency,” Journal of Visual Commu-
nication and Image Representation, vol. 4, no. 4, pp. 324–335, 1993.
[82] N. Zhao, Q. Wei, A. Basarab, N. Dobigeon, D. Kouamé, and J.-
Y. Tourneret, “Fast single image super-resolution using a new
analytical solution for `2 -`2 problems,” IEEE Transactions on Image
Processing, vol. 25, no. 8, pp. 3683–3697, 2016.
[83] K. Zhang, L. Van Gool, and R. Timofte, “Deep unfolding network
for image super-resolution,” in IEEE Conference on Computer Vision
and Pattern Recognition, 2020, pp. 3217–3226.
[84] Y. Zhang, K. Li, K. Li, L. Wang, B. Zhong, and Y. Fu, “Image super-
resolution using very deep residual channel attention networks,”
in Proceedings of the European Conference on Computer Vision, 2018,
pp. 286–301.