LDP: A Large Diffuse Filter Pruning to Slim the CNN

被引:0
作者
Wei, Wenyue [1 ]
Wang, Yizhen [1 ]
Li, Yun [1 ]
Xia, Yinfeng [1 ]
Yin, Baoqun [1 ]
机构
[1] Univ Sci & Technol China, Langfang, Peoples R China
来源
PROCEEDINGS OF 2022 THE 6TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING, ICMLSC 20222 | 2022年
关键词
Model Compression; Filter Pruning; Deactivation; Entropy Weight Calculation;
D O I
10.1145/3523150.3523155
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, filter pruning has become one of the most promising methods for CNN compression. However, most of the existing approaches need to prune filters iteratively, which increases the cost and complexity of the pruning process. In this paper, we propose a novel filter pruning strategy to conduct the removal operation only once, which avoids reconstructing new thinner models over and over. To accomplish this, we focus on the information amount of post-features, and replace the iterative removals with equally efficacious but simpler deactivation. The network repeatedly goes through filter deactivation and accuracy recovery, with more and more inefficient filters deactivated, just like a gas diffusing in a container. Therefore the proposed method is named as a Large Diffuse Pruning (LDP). Besides, we simplify the entropy weight calculation procedures by eliminating the repetitive Min-Max Normalization. Extensive experiments on VGGNet and ResNet show that our approach could achieve higher or comparable performance than some of the state-of-the-art methods.
引用
收藏
页码:26 / 32
页数:7
相关论文
共 26 条
[1]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587
[2]  
Han S, 2016, Arxiv, DOI arXiv:1510.00149
[3]  
He Y, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2234
[4]   Filter Pruning via Geometric Median for Deep Convolutional Neural Networks Acceleration [J].
He, Yang ;
Liu, Ping ;
Wang, Ziwei ;
Hu, Zhilan ;
Yang, Yi .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4335-4344
[5]   Data-Driven Sparse Structure Selection for Deep Neural Networks [J].
Huang, Zehao ;
Wang, Naiyan .
COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 :317-334
[6]  
LeCun Y., 1990, ADV NEURAL INFORM PR, V2
[7]  
Li H, 2017, Arxiv, DOI arXiv:1608.08710
[8]   Exploiting Kernel Sparsity and Entropy for Interpretable CNN Compression [J].
Li, Yuchao ;
Lin, Shaohui ;
Zhang, Baochang ;
Liu, Jianzhuang ;
Doermann, David ;
Wu, Yongjian ;
Huang, Feiyue ;
Ji, Rongrong .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :2795-2804
[9]   Using Feature Entropy to Guide Filter Pruning for Efficient Convolutional Networks [J].
Li, Yun ;
Wang, Luyang ;
Peng, Sifan ;
Kumar, Aakash ;
Yin, Baoqun .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: DEEP LEARNING, PT II, 2019, 11728 :263-274
[10]  
Lin MB, 2020, Arxiv, DOI arXiv:2001.08565