Filter pruning with a feature map entropy importance criterion for convolution neural networks compressing

被引:32
作者
Wang, Jielei [1 ]
Jiang, Ting [2 ]
Cui, Zongyong [1 ]
Cao, Zongjie [1 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu 611731, Peoples R China
[2] Megvii Technol Ltd, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; Model compression; Model pruning; Model acceleration; Entropy; GRADIENT;
D O I
10.1016/j.neucom.2021.07.034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNN) has made significant progress in recent years. However, its high computing and storage costs make it challenging to apply on resource-limited platforms or edge computation scenarios. Recent studies have shown that model pruning is an effective method to solve this problem. Typically, the model pruning method is a three-stage pipeline: training, pruning, and fine-tuning. In this work, a novel structured pruning method for Convolutional Neural Networks (CNN) compression is proposed, where filter-level redundant weights are pruned according to entropy importance criteria (termed FPEI). In short, the FPEI criterion, which works in the stage of pruning, defines the importance of the filter according to the entropy of feature maps. If a feature map contains very little information, it should not contribute much to the whole network. By removing these uninformative feature maps, their corresponding filters in the current layer and kernels in the next layer can be removed simultaneously. Consequently, the computing and storage costs are significantly reduced. Moreover, because our method cannot show the advantages of the existing ResNet pruning strategy, we propose a dimensionality reduction (DR) pruning strategy for ResNet structured networks. Experiments on several datasets demonstrate that our method is effective. In the experiment about the VGG-16 model on the SVHN dataset, we removed 91.31% of the parameters, from 14.73M to 1.28M, achieving a 63.77% reduction in the FLOPs, from 313.4M to 113.5M, and 1.73 times speedups of model inference. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:41 / 54
页数:14
相关论文
共 39 条
  • [21] HCov: A Target Attention-based Filter Pruning with Retaining High-Covariance Feature Map
    Zhang, Chenrui
    Ma, Yinan
    Wu, Jing
    Long, Chengnian
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [22] Batch-Normalization-based Soft Filter Pruning for Deep Convolutional Neural Networks
    Xu, Xiaozhou
    Chen, Qiming
    Xie, Lei
    Su, Hongye
    16TH IEEE INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2020), 2020, : 951 - 956
  • [23] Max-Plus Operators Applied to Filter Selection and Model Pruning in Neural Networks
    Zhang, Yunxiang
    Blusseau, Samy
    Velasco-Forero, Santiago
    Bloch, Isabelle
    Angulo, Jesus
    MATHEMATICAL MORPHOLOGY AND ITS APPLICATIONS TO SIGNAL AND IMAGE PROCESSING, ISMM 2019, 2019, 11564 : 310 - 322
  • [24] A Spectral Clustering Based Filter-Level Pruning Method for Convolutional Neural Networks
    Li, Lianqiang
    Zhu, Jie
    Sun, Ming-Ting
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2019, E102D (12) : 2624 - 2627
  • [25] FEATURE MAP AUGMENTATION TO IMPROVE SCALE INVARIANCE IN CONVOLUTIONAL NEURAL NETWORKS
    Kumar, Dinesh
    Sharma, Dharmendra
    JOURNAL OF ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING RESEARCH, 2023, 13 (01) : 51 - 74
  • [26] Feature Map Upscaling to Improve Scale Invariance in Convolutional Neural Networks
    Kumar, Dinesh
    Sharma, Dharmendra
    VISAPP: PROCEEDINGS OF THE 16TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS - VOL. 5: VISAPP, 2021, : 113 - 122
  • [27] Gradual Channel Pruning While Training Using Feature Relevance Scores for Convolutional Neural Networks
    Aketi, Sai Aparna
    Roy, Sourjya
    Raghunathan, Anand
    Roy, Kaushik
    IEEE ACCESS, 2020, 8 : 171924 - 171932
  • [28] Fpar: filter pruning via attention and rank enhancement for deep convolutional neural networks acceleration
    Chen, Yanming
    Wu, Gang
    Shuai, Mingrui
    Lou, Shubin
    Zhang, Yiwen
    An, Zhulin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2973 - 2985
  • [29] Accelerating Image Classification using Feature Map Similarity in Convolutional Neural Networks
    Park, Keunyoung
    Kim, Doo-Hyun
    APPLIED SCIENCES-BASEL, 2019, 9 (01):
  • [30] Elastic Filter Prune in Deep Neural Networks Using Modified Weighted Hybrid Criterion
    Hu, Wei
    Han, Yi
    Liu, Fang
    Hu, Mingce
    Li, Xingyuan
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I, KSEM 2024, 2024, 14884 : 16 - 27