Professional Documents
Culture Documents
CPFD Based On CKN
CPFD Based On CKN
CPFD Based On CKN
1 Introduction
With the rapid development of the digital image editing tools, it is easy to
tamper with the digital image even without leaving any perceptible traces.
Xianfeng Zhao
E-mail: zhaoxianfeng@iie.ac.cn
1. State Key Laboratory of Information Security, Institute of Information Engineering, Chi-
nese Academy of Sciences, Beijing 100093, China
2. School of Cyber Security, University of Chinese Academy of Sciences, Beijing 100093,
China
2 Yaqi Liu 1,2 et al.
2 Related Work
In the matching stage, similar patches will be detected. In the newly pro-
posed method in [9], PatchMatch algorithm is adopted to conduct feature
matching with high efficiency, while Zernike Moments (ZM), Polar Cosine
Transform (PCT) and Fourier-Mellin Transform (FMT) are considered for
feature extraction. After the matching stage, it is inevitable that there are
spurious pairs. The filtering stage is designed to remove those spurious pairs.
In [48], a novel filtering algorithm was proposed which can effectively prune the
falsely matched regions, besides their newly proposed interest point detector
as above mentioned.
In the post-processing stage, some simple processes, e.g., morphological
operations, are mostly employed to only preserve matches that exhibit a com-
mon behavior. To refine the results further, some newly proposed methods
update the matching information using the achieved knowledge from the pre-
vious iterations [38,19]. In [11], Ferreira et al. combined different properties
of copy-move detection approaches, and modeled the problem on a multiscale
behavior knowledge space, which encodes the output combinations of different
techniques as the priori probabilities considering multiple scales of the training
data.
Although various methods were proposed recently, leading to tremendous
progress in copy-move forgery detection, few work has been conducted on
the optimization of feature extraction. In the state-of-the-art methods, con-
ventional hand-crafted descriptors (e.g., LBP, ZM, PCT, FMT, SIFT, SURF,
etc.[8,11]) are widely adopted. Motivated by the great advance of deep learning
methods in computer vision tasks [17,35,15], we adopt a kind of data-driven
local convolutional feature, namely CKN, to conduct copy-move forgery de-
tection.
Data-driven descriptors: The algorithms adopted in copy-move forgery
detection are mainly borrowed from the computer vision tasks, such as image
classification [47], object detection [10] and image retrieval [42], etc. Recent
advances in computer vision tasks have been greatly promoted by the quick
development of the GPU technologies and the success of convolutional neural
networks (CNN) [17]. Different from conventional formulations of image clas-
sification based on local descriptors and VLAD [16] etc., the newly proposed
image classification methods based on CNN adopt an end-to-end structure.
Deep networks naturally integrate low/mid/high level features [49] and clas-
sifiers in an end-to-end multilayer fashion, and the levels of features can be
enriched by the number of stacked layers (depth). Typical convolutional neural
networks, e.g., AlexNet [17], VGG [41], ResNet [14,15], and ResNeXt [44] etc.,
have greatly improved the performance on the tasks of image classification and
object detection [35], etc. Features output by above-mentioned CNNs’ inter-
mediate layers can be regarded as image-level descriptors or so-called global
features. Those global features are designed to reinforce inter class difference
while neglect the intra class difference. And those deep learning based meth-
ods proposed for computer vision tasks can not directly be used in copy-move
forgery detection which aims to find similar regions undergoing rotation, re-
sizing or deformation.
6 Yaqi Liu 1,2 et al.
3 Method
Fig. 2 The framework of the proposed copy-move forgery detection method based on CKN.
Copy-move Forgery Detection based on Convolutional Kernel Network 7
2
X − ||u−z||
g1 (u; M) := e 2β 2
1 h1 (z; M), u ∈ Ω1 (3)
z∈Ω
2
√ − ||wjα−p̃2z || n1
h1 (z; M) := ||pz ||[ ηj e 1 ]j=1 , z ∈ Ω (4)
where Ω1 is the subset of Ω, wj and ηj are the learned parameters. There are
two distinct approximations: 1) one is in the subsampling defined by |Ω1 | ≤ |Ω|
8 Yaqi Liu 1,2 et al.
that corresponds to the stride of the pooling operation in CNN; 2) the other
is in the embedding of the Gaussian kernel of the subpatches:
||p̃z −p̃′ ′ ||2
− z
||pz ||||p′z′ ||e 2α1 2
≈ h1 (z; M)T h1 (z ′ ; M′ ) (5)
the input features are normalized, the inner part of the match kernel ||p̃z − p̃′z′ ||
is directly linked to the cosine of the angle between the two gradients, see [25].
So the approximation of kernel K1 is computed as:
||p̃z −p̃′ ′ ||2 n1
− z X
e 2α1 2
≈ ϕ1 (j; pz )ϕ1 (j; p′z′ ) (7)
j=1
where conv(·) denotes the convolutional operation, Kg (γ1 ) denotes the Gaus-
sian kernel with a factor γ1 (we set γ1 = 3), Ω1 is obtained by subsampling Ω
with the stride of γ1 . With the factor γ1 , Lk1 = 2 × γ1 + 1, the size of Kg (γ1 )
is Lk1 × Lk1 , and Kg (γ1 ) is computed as:
!
kg (k1 , k2 , γ1 )
Kg (γ1 ) = P P (13)
k1 k2 kg (k1 , k2 , γ1 ) Lk1 ×Lk1
√
where kg (k1 , k2 , γ1 ) = exp(−(k1 2 + k2 2 )/2(γ1 / 2)2 ), k1 , k2 are the relative
coordinates to the center of the kernel Kg (γ1 ). Thus, the output map of the
first layer is a tensor of the size of (m/γ1 ) × (m/γ1 ) × n1 = 17 × 17 × 16.
In the second layer, the input map is f (M), the size of the sub-patch py
is mpy × mpy = 4 × 4, subsampling factor is set as γ2 = 2 and n2 = 1024.
Omitting the borders, the number of input sub-patches is (m/γ1 − mpy + 1) ×
(m/γ1 − mpy + 1) × n1 = 14 × 14 × 16. So the input of the second layer can be
transformed to a matrix M2 of size 256×196. By conducting the approximation
n2 n2
procedure introduced in 3.1, the parameters Wi = (wj )j=1 and η i = (ηj )j=1
are learned, i = 1, · · · , N (N = 256). Trying to formulate it as the matrix
computation, the parameters are transformed to a weight matrix W1024×256
and bias B1024 , the elements are computed as:
2wj
wj = (14)
α2 2
10 Yaqi Liu 1,2 et al.
So the size of the output map is 1024 × 196. With the subsampling factor set
as γ2 = 2, the size of the final output map is 512 × 98 after linear pooling with
Gaussian weights. The computation procedure of linear pooling with Gaussian
weights is the same as formula (12). Thus, the total dimension of the feature
vector extracted by CKN-grad is 50176. Then PCA (Principle component anal-
ysis) is adopted for dimensionality reduction. The hyperparameters and the
PCA matrix both are obtained by training on the RomePatches dataset [29].
Finally, a 1024-dimensional feature vector can be extracted by CKN-grad from
a patch of size 51 × 51.
Fig. 6 The comparison between copy-move forgery detection based on SLIC and COB.
Once the keypoints Ko are detected, the CKN features can be extracted
from those detected keypoints as introduced in Section 3.2. Then the rest steps
are the same as the original work proposed in [19], which can be concluded
as following steps: (1) the detection of the suspicious pairs of regions which
contain many similar keypoints. Specifically, in each region, for each keypoint,
we search its K nearest neighbors that are located in the other regions, with
constructing a k-d tree to decrease the complexity of searching K nearest
neighbors. (2) After the step (1), suspicious pairs of regions are detected, then
we estimate the relationship between these two regions in terms of a transform
matrix by conducting RANSAC method. (3) For the reason of that a limited
number of keypoints cannot resist the possible errors in keypoint extraction,
a so-called second stage of matching is conducted to eliminate false alarm
regions. As shown in Fig. 2, this stage consists of two steps, namely obtaining
new correspondences and obtaining new transform matrix. The estimation
of the transform matrix is refined via an EM-based algorithm. Due to space
14 Yaqi Liu 1,2 et al.
4 Experiments
In this paper, the contributions are two-fold: the data-driven convolutional lo-
cal descriptor adoption, and the appropriate formulation of CKN-based copy-
move forgery detection to achieve the state-of-the-art performance, which have
been discussed in Section 1. Thus, we conduct the experiments from two as-
pects: CKN evaluation (Section 4.1) and comparison with other methods (Sec-
tion 4.2). In Section 4.1, we try to demonstrate the effectiveness and robustness
of our GPU-based CKN; In Section 4.2, we try to demonstrate that the pro-
posed method can achieve the state-of-the-art performance, and it is robust
to different kinds of attacks.
First of all, experiments are conducted with Th and min cluster pts (which
denotes the least number of pairs of matched points linking a cluster to another
one) fixed. Ward, single, and centroid are three kinds of linkage metrics used
to stop cluster grouping with the threshold (Th ), readers can refer to [2] for
details. Four kinds of methods are tested, namely, SIFT-origin, SIFT-VLFeat,
CKN-grad and CKN-grad-GPU. The results of SIFT-origin are generated by
the codes provided by the seminal work [2], the patch size of SIFT-origin is
16 × 16. For fair comparison, in SIFT-VLFeat, we extract SIFT features from
patches of size 51 × 51. In SIFT-VLFeat, CKN-grad and CKN-grad-GPU, we
adopt the codes provided by VLFeat to conduct patch extraction.
It can be seen from Table 2 that CKN-grad and CKN-grad-GPU can gen-
erate the same results without any difference, which means that there is no
significant compromising of effectiveness by reformulating CKN on GPU, while
the GPU version is more efficient. For this reason, a comprehensive comparison
is conducted in the next part among SIFT-origin, SIFT-VLFeat and CKN-
grad-GPU (we consider that CKN-grad and CKN-grad-GPU are the same).
In general, SIFT-VLFeat and CKN-grad-GPU can constantly achieve better
performance than SIFT-origin. For the reason of that the performance of SIFT
can be influenced by the patch size [39]. Besides, the number of patches gen-
erated by VLFeat is not the same as the original codes [2], though they both
adopt DoG algorithms. SIFT-VLFeat and CKN-grad-GPU can achieve similar
TPRs, while CKN-grad-GPU can achieve lower FPRs.
90 90 90
80 80 80
70 70 70
60 60 60
TPR−CKN−grad−GPU
PR (%)
PR (%)
PR (%)
50 50 50 FPR−CKN−grad−GPU
TPR−CKN−grad−GPU TPR−SIFT−origin
40 FPR−CKN−grad−GPU 40 40 FPR−SIFT−origin
TPR−SIFT−origin TPR−CKN−grad−GPU
30 FPR−SIFT−origin 30 FPR−CKN−grad−GPU 30
TPR−SIFT−origin
20 20 FPR−SIFT−origin 20
10 10 10
0 0 0
1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3
90 90 90
80 80 80
70 70 70
60 60 60
PR (%)
PR (%)
PR (%)
50 50 50
TPR−CKN−grad−GPU TPR−CKN−grad−GPU TPR−CKN−grad−GPU
40 FPR−CKN−grad−GPU 40 FPR−CKN−grad−GPU 40 FPR−CKN−grad−GPU
TPR−SIFT−origin TPR−SIFT−origin TPR−SIFT−origin
30 FPR−SIFT−origin 30 FPR−SIFT−origin 30 FPR−SIFT−origin
20 20 20
10 10 10
0 0 0
1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3
Fig. 7 The comparison between SIFT-origin and CKN-grad-GPU on MICC-F220 (top row)
and MICC-F2000 (bottom row) for different linkage metrics and Th (axis x).
90 90 90
80 80 80
70 70 70
60 60 60
PR (%)
PR (%)
PR (%)
TPR−CKN−grad−GPU
50 FPR−CKN−grad−GPU 50 50
TPR−SIFT−VLFeat TPR−CKN−grad−GPU
40 FPR−SIFT−VLFeat 40 40 FPR−CKN−grad−GPU
TPR−SIFT−VLFeat
TPR−CKN−grad−GPU
30 30 30 FPR−SIFT−VLFeat
FPR−CKN−grad−GPU
20 20 TPR−SIFT−VLFeat 20
FPR−SIFT−VLFeat
10 10 10
0 0 0
1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3
90 90 90
80 80
80
70 70
70 TPR−CKN−grad−GPU
60 60 FPR−CKN−grad−GPU
60 TPR−SIFT−VLFeat
PR (%)
PR (%)
PR (%)
TPR−CKN−grad−GPU 50 50 FPR−SIFT−VLFeat
50 FPR−CKN−grad−GPU
TPR−SIFT−VLFeat 40 40
FPR−SIFT−VLFeat TPR−CKN−grad−GPU
40 FPR−CKN−grad−GPU
30 30
TPR−SIFT−VLFeat
30 FPR−SIFT−VLFeat
20 20
20 10 10
10 0 0
1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3 1.6 1.8 2 2.2 2.4 2.6 2.8 3
5 Conclusion
Table 5 Copy-move forgery detection results on CoMoFoD dataset with JPEG compression
(quality factor = 90).
Acknowledgements This work was supported by the NSFC under U1636102 and U1536105,
and National Key Technology R&D Program under 2014BAH41B01, 2016YFB0801003 and
2016QY15Z2500.
Copy-move Forgery Detection based on Convolutional Kernel Network 21
Table 6 Copy-move forgery detection results on CoMoFoD dataset with Noise adding (vari-
ance = 0.0005).
Table 7 Copy-move forgery detection results on CoMoFoD dataset with image blurring
(averaging filter = 3 × 3).
References
1. Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., Süsstrunk, S.: Slic superpixels
compared to state-of-the-art superpixel methods. IEEE transactions on pattern analysis
and machine intelligence 34(11), 2274–2282 (2012)
2. Amerini, I., Ballan, L., Caldelli, R., Bimbo, A.D., Serra, G.: A sift-based forensic method
for copycmove attack detection and transformation recovery. IEEE Transactions on
Information Forensics and Security 6(3), 1099–1110 (2011)
3. Arbelaez, P., Maire, M., Fowlkes, C., Malik, J.: Contour detection and hierarchical image
segmentation. IEEE transactions on pattern analysis and machine intelligence 33(5),
898–916 (2011)
22 Yaqi Liu 1,2 et al.
Table 8 Copy-move forgery detection results on CoMoFoD dataset with brightness change
((lower bound, upper bound) = (0.01, 0.8)).
Table 9 Copy-move forgery detection results on CoMoFoD dataset with color reduction
(intensity levels per each color channel = 32).
4. Ardizzone, E., Bruno, A., Mazzola, G.: Copycmove forgery detection by matching tri-
angles of keypoints. IEEE Transactions on Information Forensics and Security 10(10),
2084–2094 (2015)
5. Bashar, M., Noda, K., Ohnishi, N., Mori, K.: Exploring duplicated regions in natural
images. IEEE Transactions on Image Processing (2010)
6. Bayar, B., Stamm, M.C.: A deep learning approach to universal image manipulation
detection using a new convolutional layer. In: Proceedings of the 4th ACM Workshop
on Information Hiding and Multimedia Security, pp. 5–10. ACM (2016)
7. Borji, A., Cheng, M.M., Jiang, H., Li, J.: Salient object detection: A survey. arXiv
preprint arXiv:1411.5878 (2014)
Copy-move Forgery Detection based on Convolutional Kernel Network 23
Table 10 Copy-move forgery detection results on CoMoFoD dataset with contrast adjust-
ments ((lower bound, upper bound) = (0.01, 0.8)).
8. Christlein, V., Riess, C., Jordan, J., Riess, C., Angelopoulou, E.: An evaluation of
popular copy-move forgery detection approaches. IEEE Transactions on Information
Forensics and Security 7(6), 1841–1854 (2012)
9. Cozzolino, D., Poggi, G., Verdoliva, L.: Efficient dense-field copycmove forgery detection.
IEEE Transactions on Information Forensics and Security 10(11), 2284–2297 (2015)
10. Felzenszwalb, P.F., Girshick, R.B., McAllester, D., Ramanan, D.: Object detection with
discriminatively trained part-based models. IEEE Transactions on Pattern Analysis and
Machine Intelligence 32(9), 1627–1645 (2010)
11. Ferreira, A., Felipussi, S.C., Alfaro, C., Fonseca, P., Vargas-Munoz, J.E., dos Santos,
J.A., Rocha, A.: Behavior knowledge space-based fusion for copycmove forgery detec-
tion. IEEE Transactions on Image Processing 25(10), 4729–4742 (2016)
12. Fischer, P., Dosovitskiy, A., Brox, T.: Descriptor matching with convolutional neural
networks: a comparison to sift. In: arXiv preprint arXiv:1405.5769 (2014)
13. Fridrich, A.J., Soukal, B.D., Lukáš, A.J.: Detection of copy-move forgery in digital
images. In: in Proceedings of Digital Forensic Research Workshop. Citeseer (2003)
14. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In:
arXiv preprint arXiv:1512.03385 (2015)
15. He, K., Zhang, X., Ren, S., Sun, J.: Identity mappings in deep residual networks. In:
arXiv preprint arXiv:1603.05027 (2016)
16. Jegou, H., Perronnin, F., Douze, M., Sanchez, J., Perez, P., Schmid, C.: Aggregating
local image descriptors into compact codes. IEEE Transactions on Pattern Analysis
and Machine Intelligence 34(9), 1704–1716 (2012)
17. Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolu-
tional neural networks. In: Proceedings of the Advances in neural information processing
systems, pp. 1097–1105 (2012)
18. Li, G., Wu, Q., Tu, D., Sun, S.: A sorted neighborhood approach for detecting duplicated
regions in image forgeries based on dwt and svd. In: Multimedia and Expo, 2007 IEEE
International Conference on, pp. 1750–1753. IEEE (2007)
19. Li, J., Li, X., Yang, B., Sun, X.: Segmentation-based image copy-move forgery detec-
tion scheme. IEEE Transactions on Information Forensics and Security 10(3), 507–518
(2015)
20. Li, L., Li, S., Zhu, H., Chu, S.C., Roddick, J.F., Pan, J.S.: An efficient scheme for
detecting copy-move forged images by local binary patterns. Journal of Information
Hiding and Multimedia Signal Processing 4(1), 46–56 (2013)
24 Yaqi Liu 1,2 et al.
21. Li, Y.: Image copy-move forgery detection based on polar cosine transform and approx-
imate nearest neighbor searching. Forensic science international 224(1), 59–67 (2013)
22. Liu, Y., Cai, Q., Zhu, X., Cao, J., Li, H.: Saliency detection using two-stage scoring.
In: Image Processing (ICIP), 2015 IEEE International Conference on, pp. 4062–4066.
IEEE (2015)
23. Mahdian, B., Saic, S.: Detection of copy–move forgery using a method based on blur
moment invariants. Forensic science international 171(2), 180–189 (2007)
24. Mairal, J.: End-to-end kernel learning with supervised convolutional kernel networks.
In: Proceedings of the Advances in neural information processing systems (2016)
25. Mairal, J., Koniusz, P., Harchaoui, Z., Schmid, C.: Convolutional kernel networks. In:
Proceedings of the Advances in neural information processing systems, pp. 2627–2635
(2014)
26. Maninis, K.K., Pont-Tuset, J., Arbeláez, P., Van Gool, L.: Convolutional oriented
boundaries. In: European Conference on Computer Vision, pp. 580–596. Springer (2016)
27. Mostajabi, M., Yadollahpour, P., Shakhnarovich, G.: Feedforward semantic segmenta-
tion with zoom-out features. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 3376–3385 (2015)
28. Pan, X., Lyu, S.: Region duplication detection using image feature matching. IEEE
Transactions on Information Forensics and Security 5(4), 857–867 (2010)
29. Paulin, M., Douze, M., Harchaoui, Z., Mairal, J., Perronin, F., Schmid, C.: Local con-
volutional features with unsupervised training for image retrieval. In: Proceedings of
the International Conference on Computer Vision, pp. 91–99. IEEE (2015)
30. Paulin, M., Mairal, J., Douze, M., Harchaoui, Z., Perronnin, F., Schmid, C.: Convolu-
tional patch representations for image retrieval: an unsupervised approach. International
Journal of Computer Vision 121(1), 149–168 (2017)
31. Pont-Tuset, J., Arbelaez, P., Barron, J.T., Marques, F., Malik, J.: Multiscale combinato-
rial grouping for image segmentation and object proposal generation. IEEE transactions
on pattern analysis and machine intelligence 39(1), 128–140 (2017)
32. Popescu, A., Farid, H.: Exposing digital forgeries by detecting duplicated image region
[technical report]. 2004-515. Hanover, Department of Computer Science, Dartmouth
College. USA p. 32 (2004)
33. Pun, C.M., Yuan, X.C., Bi, X.L.: Image forgery detection using adaptive oversegmen-
tation and feature point matching. IEEE Transactions on Information Forensics and
Security 10(8), 1705–1716 (2015)
34. Rao, Y., Ni, J.: A deep learning approach to detection of splicing and copy-move forgeries
in images. In: IEEE International Workshop on Information Forensics and Security
(WIFS), pp. 1–6. IEEE (2016)
35. Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: Towards real-time object detection
with region proposal networks. In: Proceedings of the Advances in neural information
processing systems, pp. 91–99 (2015)
36. Ryu, S.J., Kirchner, M., Lee, M.J., Lee, H.K.: Rotation invariant localization of du-
plicated image regions based on zernike moments. IEEE Transactions on Information
Forensics and Security 8(8), 1355–1370 (2013)
37. Shivakumar, B., Baboo, L.D.S.S.: Detection of region duplication forgery in digital
images using surf. IJCSI International Journal of Computer Science Issues 8(4) (2011)
38. Silva, E., Carvalho, T., Ferreira, A., Rocha, A.: Going deeper into copy-move forgery de-
tection: Exploring image telltales via multi-scale analysis and voting processes. Journal
of Visual Communication and Image Representation 29, 16–C32 (2015)
39. Simo-Serra, E., Trulls, E., Ferraz, L., Kokkinos, I., Fua, P., Moreno-Noguer, F.: Dis-
criminative learning of deep convolutional feature point descriptors. In: Proceedings of
the IEEE International Conference on Computer Vision, pp. 118–126 (2015)
40. Simo-Serra, E., Trulls, E., Ferraz, L., Kokkinos, I., Moreno-Noguer, F.: Fracking deep
convolutional image descriptors. In: arXiv preprint arXiv:1412.6537 (2014)
41. Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image
recognition. In: arXiv preprint arXiv:1409.155 (2015)
42. Smeulders, A.W., Worring, M., Santini, S., Gupta, A., Jain, R.: Content-based image
retrieval at the end of the early years. IEEE Transactions on Pattern Analysis and
Machine Intelligence 22(12), 1349–1380 (2000)
Copy-move Forgery Detection based on Convolutional Kernel Network 25
43. Tralic, D., Zupancic, I., Grgic, S., Grgic, M.: Comofodnew database for copy-move
forgery detection. In: ELMAR, 2013 55th international symposium, pp. 49–54. IEEE
(2013)
44. Xie, S., Girshick, R., Dollar, P., Tu, Z., He, K.: Aggregated residual transformations for
deep neural networks. In: arXiv preprint arXiv:1611.05431 (2016)
45. Xie, S., Tu, Z.: Holistically-nested edge detection. In: Proceedings of the IEEE Inter-
national Conference on Computer Vision, pp. 1395–1403 (2015)
46. Yang, B., Sun, X., Guo, H., Xia, Z., Chen, X.: A copy-move forgery detection method
based on cmfd-sift. Multimedia Tools and Applications pp. 1–19 (2017)
47. Yang, J., Yu, K., Gong, Y., Huang, T.: Linear spatial pyramid matching using sparse
coding for image classification. In: Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, pp. 20–25. IEEE (2009)
48. Zandi, M., Mahmoudi-Aznaveh, A., Talebpour, A.: Iterative copy-move forgery detec-
tion based on a new interest point detector. IEEE Transactions on Information Forensics
and Security 11(11), 2499–2512 (2016)
49. Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In:
Proceedings of the European Conference on Computer Vision, pp. 818–833. Springer
(2014)