Filter pruning with a feature map entropy importance criterion for convolution neural networks compressing

被引:37
作者
Wang, Jielei [1 ]
Jiang, Ting [2 ]
Cui, Zongyong [1 ]
Cao, Zongjie [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 611731, Peoples R China
[2] Megvii Technol Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Model compression; Model pruning; Model acceleration; Entropy; GRADIENT;
D O I
10.1016/j.neucom.2021.07.034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNN) has made significant progress in recent years. However, its high computing and storage costs make it challenging to apply on resource-limited platforms or edge computation scenarios. Recent studies have shown that model pruning is an effective method to solve this problem. Typically, the model pruning method is a three-stage pipeline: training, pruning, and fine-tuning. In this work, a novel structured pruning method for Convolutional Neural Networks (CNN) compression is proposed, where filter-level redundant weights are pruned according to entropy importance criteria (termed FPEI). In short, the FPEI criterion, which works in the stage of pruning, defines the importance of the filter according to the entropy of feature maps. If a feature map contains very little information, it should not contribute much to the whole network. By removing these uninformative feature maps, their corresponding filters in the current layer and kernels in the next layer can be removed simultaneously. Consequently, the computing and storage costs are significantly reduced. Moreover, because our method cannot show the advantages of the existing ResNet pruning strategy, we propose a dimensionality reduction (DR) pruning strategy for ResNet structured networks. Experiments on several datasets demonstrate that our method is effective. In the experiment about the VGG-16 model on the SVHN dataset, we removed 91.31% of the parameters, from 14.73M to 1.28M, achieving a 63.77% reduction in the FLOPs, from 313.4M to 113.5M, and 1.73 times speedups of model inference. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:41 / 54
页数:14
相关论文
共 40 条
[31]   Elastic Filter Prune in Deep Neural Networks Using Modified Weighted Hybrid Criterion [J].
Hu, Wei ;
Han, Yi ;
Liu, Fang ;
Hu, Mingce ;
Li, Xingyuan .
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2024, 2024, 14884 :16-27
[32]   Compression of Deep Convolutional Neural Network Using Additional Importance-Weight-Based Filter Pruning Approach [J].
Sawant, Shrutika S. ;
Wiedmann, Marco ;
Goeb, Stephan ;
Holzer, Nina ;
Lang, Elmar W. ;
Goetz, Theresa .
APPLIED SCIENCES-BASEL, 2022, 12 (21)
[33]   Application of Feature Extraction through Convolution Neural Networks and SVM Classifier for Robust Grading of Apples [J].
Yuan CAI ;
Clarence WDE SILVA ;
Bing LI ;
Liqun WANG ;
Ziwen WANG .
Instrumentation, 2019, 6 (04) :59-71
[34]   Optimization assisted autoregressive technique with deep convolution neural network-based entropy filter for image demosaicing [J].
Mary, C. Anitha ;
Wesley, A. Boyed .
IMAGING SCIENCE JOURNAL, 2024, 72 (08) :1108-1125
[35]   Batch Entropy Supervised Convolutional Neural Networks for Feature Extraction and Harmonizing for Action Recognition [J].
Hossain, Md Imtiaz ;
Siddique, Ashraf ;
Hossain, Md Alamgir ;
Hossain, Md Delowar ;
Huh, Eui-Nam .
IEEE ACCESS, 2020, 8 :206427-206444
[36]   Pruning Artificial Neural Networks: A Way to Find Well-Generalizing, High-Entropy Sharp Minima [J].
Tartaglione, Enzo ;
Bragagnolo, Andrea ;
Grangetto, Marco .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 :67-78
[37]   Auto-Tiler: Variable-Dimension Autoencoder with Tiling for Compressing Intermediate Feature Space of Deep Neural Networks for Internet of Things [J].
Park, Jeongsoo ;
Kim, Jungrae ;
Ko, Jong Hwan .
SENSORS, 2021, 21 (03) :1-17
[38]   Mitigating severe over-parameterization in deep convolutional neural networks through forced feature abstraction and compression with an entropy-based heuristic [J].
Gowdra, Nidhi ;
Sinha, Roopak ;
MacDonell, Stephen ;
Yan, Wei Qi .
PATTERN RECOGNITION, 2021, 119
[39]   DC-AC: Deep Correlation-based Adaptive Compression of Feature Map Planes in Convolutional Neural Networks [J].
Bae, Seung-Hwan ;
Lee, Hyuk-Jae ;
Kim, Hyun .
2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
[40]   Total contribution score and fuzzy entropy based two-stage selection of FC, ReLU and inverseReLU features of multiple convolution neural networks for erythrocytes detection [J].
Banerjee, Sriparna ;
Chaudhuri, Sheli Sinha .
IET COMPUTER VISION, 2019, 13 (07) :640-650