Automatic Compression Ratio Allocation for Pruning Convolutional Neural Networks

被引:0
|
作者
Liu, Yunfeng [1 ]
Kong, Huihui [1 ]
Yu, Peihua [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
来源
ICVISP 2019: PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING | 2019年
基金
中国国家自然科学基金;
关键词
Neural Networks; Network Pruning; Model Compression; Computer Vision;
D O I
10.1145/3387168.3387184
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) have demonstrated significant performance improvement in many application scenarios. However, the high computational complexity and model size have limited its application on the mobile and embedded devices. Various approaches have been proposed to compress CNNs. Filter pruning is widely considered as a promising solution, which can significantly speed up the inference and reduce memory consumption. To this end, most approaches tend to prune filters by manually allocating compression ratio, which highly relies on individual expertise and not friendly to non-professional users. In this paper, we propose an Automatic Compression Ratio Allocation (ACRA) scheme based on binary search algorithm to prune convolutional neural networks. Specifically, ACRA provides two strategies for allocating compression ratio automatically. First, uniform pruning strategy allocates the same compression ratio to each layer, which is obtained by binary search based on target FLOPs reduction of the whole networks. Second, sensitivity-based pruning strategy allocates appropriate compression ratio to each layer based on the sensitivity to accuracy. Experimental results from VGG11 and VGG-16, demonstrate that our scheme can reduce FLOPs significantly while maintaining a high accuracy level. Specifically, for the VGG16 on CIFAR-10 dataset, we reduce 29.18% FLOPs with only 1.24% accuracy decrease.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Batch-Normalization-based Soft Filter Pruning for Deep Convolutional Neural Networks
    Xu, Xiaozhou
    Chen, Qiming
    Xie, Lei
    Su, Hongye
    16TH IEEE INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2020), 2020, : 951 - 956
  • [22] Overview of Deep Convolutional Neural Network Pruning
    Li, Guang
    Liu, Fang
    Xia, Yuping
    2020 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING AND ARTIFICIAL INTELLIGENCE, 2020, 11584
  • [23] Structured Pruning for Efficient Convolutional Neural Networks via Incremental Regularization
    Wang, Huan
    Hu, Xinyi
    Zhang, Qiming
    Wang, Yuehai
    Yu, Lu
    Hu, Haoji
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 775 - 788
  • [24] Acceleration of Deep Convolutional Neural Networks Using Adaptive Filter Pruning
    Singh, Pravendra
    Verma, Vinay Kumar
    Rai, Piyush
    Namboodiri, Vinay P.
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 838 - 847
  • [25] Studying the plasticity in deep convolutional neural networks using random pruning
    Mittal, Deepak
    Bhardwaj, Shweta
    Khapra, Mitesh M.
    Ravindran, Balaraman
    MACHINE VISION AND APPLICATIONS, 2019, 30 (02) : 203 - 216
  • [26] Recursive least squares method for training and pruning convolutional neural networks
    Tianzong Yu
    Chunyuan Zhang
    Meng Ma
    Yuan Wang
    Applied Intelligence, 2023, 53 : 24603 - 24618
  • [27] Recursive least squares method for training and pruning convolutional neural networks
    Yu, Tianzong
    Zhang, Chunyuan
    Ma, Meng
    Wang, Yuan
    APPLIED INTELLIGENCE, 2023, 53 (20) : 24603 - 24618
  • [28] Studying the plasticity in deep convolutional neural networks using random pruning
    Deepak Mittal
    Shweta Bhardwaj
    Mitesh M. Khapra
    Balaraman Ravindran
    Machine Vision and Applications, 2019, 30 : 203 - 216
  • [29] Gradual Channel Pruning While Training Using Feature Relevance Scores for Convolutional Neural Networks
    Aketi, Sai Aparna
    Roy, Sourjya
    Raghunathan, Anand
    Roy, Kaushik
    IEEE ACCESS, 2020, 8 : 171924 - 171932
  • [30] Fast and Robust Compression of Deep Convolutional Neural Networks
    Wen, Jia
    Yang, Liu
    Shen, Chenyang
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT II, 2020, 12397 : 52 - 63