References

You might also like

Download as odt, pdf, or txt
Download as odt, pdf, or txt
You are on page 1of 4

\section{INTRODUCTION}

In recent years, the field of deep learning has witnessed remarkable


advancements, particularly in the domains of image recognition and retrieval.
Convolutional neural networks (CNNs) have been at the forefront of this
progress, consistently achieving impressive performance. However, the
computational demands of CNNs pose significant challenges when deploying them on
resource-limited devices. To address this, network compression techniques—
specifically pruning methods—have emerged as a solution to reduce the
computational complexity of CNNs by eliminating redundant structures.

Challenges with Existing Pruning Methods:

1. Weight-Based Pruning: Most state-of-the-art pruning techniques rank filters


based solely on their weights, without considering their actual impact on
output. This approach can inadvertently lead to the removal of critical filters
and make it challenging to determine the optimal pruning threshold.

2.Forward or Isolated Pruning: Existing methods often prune filters either in


the forward direction or in isolation, which can result in accuracy loss that is
difficult to control during the pruning process.
% Convolutional neural network(CNNs) have emerged as a crutial tool in numerous
fields such as Computer vision and natural language processing, consistently
achieving higher accuracies.

@article{geng2022pruning,
title={Pruning convolutional neural networks via filter similarity analysis},
author={Geng, Lili and Niu, Baoning},
journal={Machine Learning},
volume={111},
number={9},
pages={3161--3180},
year={2022},
publisher={Springer}
}

@article{10.1007/s10489-022-03229-5,
author = {Sawant, Shrutika S. and Bauer, J. and Erick, F. X. and Ingaleshwar,
Subodh and Holzer, N. and Ramming, A. and Lang, E. W. and G\"{o}tz, Th.},
title = {An optimal-score-based filter pruning for deep convolutional neural
networks},
year = {2022},
issue_date = {Dec 2022},
publisher = {Kluwer Academic Publishers},
address = {USA},
volume = {52},
number = {15},
issn = {0924-669X},
url = {https://doi.org/10.1007/s10489-022-03229-5},
doi = {10.1007/s10489-022-03229-5},
abstract = {Convolutional Neural Networks (CNN) have achieved excellent
performance in the processing of high-resolution images. Most of these networks
contain many deep layers in quest of greater segmentation performance. However,
over-sized CNN models result in overwhelming memory usage and large inference
costs. Earlier studies have revealed that over-sized deep neural models tend to
deal with abundant redundant filters that are very similar and provide tiny or
no contribution in accelerating the inference of the model. Therefore, we have
proposed a novel optimal-score-based filter pruning (OSFP) approach to prune
redundant filters according to their relative similarity in feature space. OSFP
not only speeds up learning in the network but also eradicates redundant filters
leading to improvement in the segmentation performance. We empirically
demonstrate on widely used segmentation network models (TernausNet, classical U-
Net and VGG16 U-Net) and benchmark datasets (Inria Aerial Image Labeling Dataset
and Aerial Imagery for Roof Segmentation (AIRS)) that computation costs (in
terms of Float Point Operations (FLOPs) and parameters) are reduced
significantly, while maintaining or even improving accuracy.},
journal = {Applied Intelligence},
month = {dec},
pages = {17557–17579},
numpages = {23},
keywords = {Redundancy, Model compression, Image segmentation, Filter pruning,
Deep learning, CNN}
}

@article{fernandes2021pruning,

title={Pruning deep convolutional neural networks architectures with evolution


strategy},
author={Fernandes Jr, Francisco E and Yen, Gary G},
journal={Information Sciences},
volume={552},
pages={29--47},
year={2021},
publisher={Elsevier}
}

@Article{app12094541,
AUTHOR = {Wu, Zhihong and Li, Fuxiang and Zhu, Yuan and Lu, Ke and Wu, Mingzhi and
Zhang, Changze},
TITLE = {A Filter Pruning Method of CNN Models Based on Feature Maps Clustering},
JOURNAL = {Applied Sciences},
VOLUME = {12},
YEAR = {2022},
NUMBER = {9},
ARTICLE-NUMBER = {4541},
URL = {https://www.mdpi.com/2076-3417/12/9/4541},
ISSN = {2076-3417},
ABSTRACT = {The convolutional neural network (CNN) has been widely used in the field of self-
driving cars. To satisfy the increasing demand, the deeper and wider neural network has become a
general trend. However, this leads to the main problem that the deep neural network is
computationally expensive and consumes a considerable amount of memory. To compress and
accelerate the deep neural network, this paper proposes a filter pruning method based on feature
maps clustering. The basic idea is that by clustering, one can know how many features the input
images have and how many filters are enough to extract all features. This paper chooses Retinanet
and WIDER FACE datasets to experiment with the proposed method. Experiments demonstrate that
the hierarchical clustering algorithm is an effective method for filtering pruning, and the silhouette
coefficient method can be used to determine the number of pruned filters. This work evaluates the
performance change by increasing the pruning ratio. The main results are as follows: Firstly, it is
effective to select pruned filters based on feature maps clustering, and its precision is higher than
that of a random selection of pruned filters. Secondly, the silhouette coefficient method is a feasible
method for finding the best clustering number. Thirdly, the detection speed of the pruned model
improves greatly. Lastly, the method we propose can be used not only for Retinanet, but also for
other CNN models. Its effect will be verified in future work.},
DOI = {10.3390/app12094541}
}

@article{zu2023consecutive,
title={Consecutive layer collaborative filter similarity for differentiable
neural network pruning},
author={Zu, Xuan and Li, Yun and Yin, Baoqun},
journal={Neurocomputing},
volume={533},
pages={35--45},
year={2023},
publisher={Elsevier}
}

@inproceedings{song2022filter,
title={Filter Pruning via Similarity Clustering for Deep Convolutional Neural
Networks},
author={Song, Kang and Yao, Wangshu and Zhu, Xuan},
booktitle={International Conference on Neural Information Processing},
pages={88--99},
year={2022},
organization={Springer}
}

@article{geng2022pruning,
title={Pruning convolutional neural networks via filter similarity analysis},
author={Geng, Lili and Niu, Baoning},
journal={Machine Learning},
volume={111},
number={9},
pages={3161--3180},
year={2022},
publisher={Springer}
}

@article{liu2023exploiting,
title={Exploiting similarity-induced redundancies in correlation topology for
channel pruning in deep convolutional neural networks},
author={Liu, J and Shao, HJ and Deng, X and Jiang, YT},
journal={International Journal of Computers and Applications},
pages={1--12},
year={2023},
publisher={Taylor \& Francis}
}

@article{vadera2022methods,
title={Methods for pruning deep neural networks},
author={Vadera, Sunil and Ameen, Salem},
journal={IEEE Access},
volume={10},
pages={63280--63300},
year={2022},
publisher={IEEE}
}

@article{molchanov2016pruning,
title={Pruning convolutional neural networks for resource efficient
inference},
author={Molchanov, Pavlo and Tyree, Stephen and Karras, Tero and Aila, Timo
and Kautz, Jan},
journal={arXiv preprint arXiv:1611.06440},
year={2016}
}

@article{yeom2021pruning,
title={Pruning by explaining: A novel criterion for deep neural network
pruning},
author={Yeom, Seul-Ki and Seegerer, Philipp and Lapuschkin, Sebastian and
Binder, Alexander and Wiedemann, Simon and M{\"u}ller, Klaus-Robert and Samek,
Wojciech},
journal={Pattern Recognition},
volume={115},
pages={107899},
year={2021},
publisher={Elsevier}
}

You might also like