Automatic Compression Ratio Allocation for Pruning Convolutional Neural Networks

被引:0
|
作者
Liu, Yunfeng [1 ]
Kong, Huihui [1 ]
Yu, Peihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
来源
ICVISP 2019: PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING | 2019年
基金
中国国家自然科学基金;
关键词
Neural Networks; Network Pruning; Model Compression; Computer Vision;
D O I
10.1145/3387168.3387184
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) have demonstrated significant performance improvement in many application scenarios. However, the high computational complexity and model size have limited its application on the mobile and embedded devices. Various approaches have been proposed to compress CNNs. Filter pruning is widely considered as a promising solution, which can significantly speed up the inference and reduce memory consumption. To this end, most approaches tend to prune filters by manually allocating compression ratio, which highly relies on individual expertise and not friendly to non-professional users. In this paper, we propose an Automatic Compression Ratio Allocation (ACRA) scheme based on binary search algorithm to prune convolutional neural networks. Specifically, ACRA provides two strategies for allocating compression ratio automatically. First, uniform pruning strategy allocates the same compression ratio to each layer, which is obtained by binary search based on target FLOPs reduction of the whole networks. Second, sensitivity-based pruning strategy allocates appropriate compression ratio to each layer based on the sensitivity to accuracy. Experimental results from VGG11 and VGG-16, demonstrate that our scheme can reduce FLOPs significantly while maintaining a high accuracy level. Specifically, for the VGG16 on CIFAR-10 dataset, we reduce 29.18% FLOPs with only 1.24% accuracy decrease.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] An Approach to Pruning the Structure of Convolutional Neural Networks without Loss of Generalization Ability
    Chen, Chaoxiang
    Kroshchanka, Aliaksandr
    Golovko, Vladimir
    Golovko, Olha
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2024, 34 (02) : 258 - 265
  • [42] RFPruning: A retraining-free pruning method for accelerating convolutional neural networks
    Wang, Zhenyu
    Xie, Xuemei
    Shi, Guangming
    APPLIED SOFT COMPUTING, 2021, 113
  • [43] Structured pruning via feature channels similarity and mutual learning for convolutional neural network compression
    Wei Yang
    Yancai Xiao
    Applied Intelligence, 2022, 52 : 14560 - 14570
  • [44] Automatic Detection of Epileptic Seizures with Recurrent and Convolutional Neural Networks
    Carrion, Salvador
    Lopez-Chilet, Alvaro
    Martinez-Bernia, Javier
    Coll-Alonso, Joan
    Chorro-Juan, Daniel
    Ander Gomez, Jon
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022 WORKSHOPS, PT I, 2022, 13373 : 522 - 532
  • [45] Structured pruning via feature channels similarity and mutual learning for convolutional neural network compression
    Yang, Wei
    Xiao, Yancai
    APPLIED INTELLIGENCE, 2022, 52 (12) : 14560 - 14570
  • [46] Activation-Based Pruning of Neural Networks
    Ganguli, Tushar
    Chong, Edwin K. P.
    Werner, Frank
    ALGORITHMS, 2024, 17 (01)
  • [47] A novel and efficient model pruning method for deep convolutional neural networks by evaluating the direct and indirect effects of filters
    Zheng, Yongbin
    Sun, Peng
    Ren, Qian
    Xu, Wanying
    Zhu, Di
    NEUROCOMPUTING, 2024, 569
  • [48] A One-step Pruning-recovery Framework for Acceleration of Convolutional Neural Networks
    Wang, Dong
    Bai, Xiao
    Zhou, Lei
    Zhou, Jun
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 768 - 775
  • [49] Data-Efficient Adaptive Global Pruning for Convolutional Neural Networks in Edge Computing
    Gao, Zhipeng
    Sun, Shan
    Mo, Zijia
    Rui, Lanlan
    Yang, Yang
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 6633 - 6638
  • [50] Filter pruning by image channel reduction in pre-trained convolutional neural networks
    Gi Su Chung
    Chee Sun Won
    Multimedia Tools and Applications, 2021, 80 : 30817 - 30826