A Study of Filter Duplication for CNNs Filter Pruning

被引:0
作者
Ikuta, Ryosuke [1 ]
Yata, Noriko [1 ]
Manabe, Yoshitsugu [1 ]
机构
[1] Chiba Univ, 1-33 Yayoicho,Inage Ku, Chiba, Chiba 2638522, Japan
来源
INTERNATIONAL WORKSHOP ON ADVANCED IMAGING TECHNOLOGY, IWAIT 2024 | 2024年 / 13164卷
关键词
CNN; pruning; redundancy; filter duplication;
D O I
10.1117/12.3018876
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional Neural Networks (CNNs) have demonstrated great success in image recognition, but most trained models are over-parameterized, and models can be compressed with only a slight performance degradation. Pruning is one of the lightweight techniques of networks, which obtains a model with a lower computational cost of inference by removing filters selectively that do not contribute to the performance. While various methods have been proposed to identify unimportant filters, determining the number of filters to be removed at each layer without causing a significant loss of accuracy is an open problem. This paper proposes a "filter duplication" approach to reduce the accuracy degradation caused by pruning, especially in higher compression ratio ranges. Filter duplication replaces unimportant filters with critical filters in a pre-trained model based on the measured importance of each convolutional layer before pruning. In experiments using mainstream CNN models and datasets, we confirmed that filter duplication improves the accuracy of the pruned model, especially with higher compression ratios. In addition, the proposed method can reflect the structural redundancy of the network to the compression ratio of each layer, providing a more efficient compression. The results show that duplicating an appropriate number of critical filters for each layer improves the robustness of the network against pruning, and optimization of duplication methods is desirable.
引用
收藏
页数:6
相关论文
共 9 条
  • [1] EIE: Efficient Inference Engine on Compressed Deep Neural Network
    Han, Song
    Liu, Xingyu
    Mao, Huizi
    Pu, Jing
    Pedram, Ardavan
    Horowitz, Mark A.
    Dally, William J.
    [J]. 2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, : 243 - 254
  • [2] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [3] Krizhevsky A., 2009, THESIS
  • [4] Kuzmin A, 2024, Arxiv, DOI arXiv:2307.02973
  • [5] Li Hao, 2017, INT C LEARN REPR ICL
  • [6] HRank: Filter Pruning using High-Rank Feature Map
    Lin, Mingbao
    Ji, Rongrong
    Wang, Yan
    Zhang, Yichen
    Zhang, Baochang
    Tian, Yonghong
    Shao, Ling
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 1526 - 1535
  • [7] Pravendra S., 2019, INT JOINT C ART INT
  • [8] Simonyan K, 2015, Arxiv, DOI [arXiv:1409.1556, 10.48550/arXiv.1409.1556,CoRR]
  • [9] Convolutional Neural Network Pruning with Structural Redundancy Reduction
    Wang, Zi
    Li, Chengcheng
    Wang, Xiangyang
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14908 - 14917