Pruning Ratio Optimization with Layer-Wise Pruning Method for Accelerating Convolutional Neural Networks

被引:3
作者
Kamma, Koji [1 ]
Inoue, Sarimu [2 ]
Wada, Toshikazu [1 ]
机构
[1] Wakayama Univ, Fac Syst Engn, Wakayama 6408510, Japan
[2] Wakayama Univ, Wakayama 6408510, Japan
关键词
pruning; pruning ratio optimizer; PRO;
D O I
10.1587/transinf.2021EDP7096
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Pruning is an effective technique to reduce computational complexity of Convolutional Neural Networks (CNNs) by removing redundant neurons (or weights). There are two types of pruning methods: holistic pruning and layer-wise pruning. The former selects the least important neuron from the entire model and prunes it. The latter conducts pruning layer by layer. Recently, it has turned out that some layer-wise methods are effective for reducing computational complexity of pruned models while preserving their accuracy. The difficulty of layer-wise pruning is how to adjust pruning ratio (the ratio of neurons to be pruned) in each layer. Because CNNs typically have lots of layers composed of lots of neurons, it is inefficient to tune pruning ratios by human hands. In this paper, we present Pruning Ratio Optimizer (PRO), a method that can be combined with layer-wise pruning methods for optimizing pruning ratios. The idea of PRO is to adjust pruning ratios based on how much pruning in each layer has an impact on the outputs in the final layer. In the experiments, we could verify the effectiveness of PRO.
引用
收藏
页码:161 / 169
页数:9
相关论文
共 31 条
[21]   SBNet: Sparse Blocks Network for Fast Inference [J].
Ren, Mengye ;
Pokrovsky, Andrei ;
Yang, Bin ;
Urtasun, Raquel .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8711-8720
[22]  
Simonyan K., P INT C LEARN REPR I
[23]  
Theis Lucas, 2018, CoRR
[24]  
Tianxing He, 2014, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), P245, DOI 10.1109/ICASSP.2014.6853595
[25]   Towards Fine-grained Network Security Forensics and Diagnosis in the SDN Era [J].
Wang, Haopei ;
Yang, Guangliang ;
Chinprutthiwong, Phakpoom ;
Xu, Lei ;
Zhang, Yangyong ;
Gu, Guofei .
PROCEEDINGS OF THE 2018 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY (CCS'18), 2018, :3-16
[26]   Interleaved Structured Sparse Convolutional Neural Networks [J].
Xie, Guotian ;
Wang, Jingdong ;
Zhang, Ting ;
Lai, Jianhuang ;
Hong, Richang ;
Qi, Guo-Jun .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8847-8856
[27]  
Xue J, 2013, INTERSPEECH, P2364
[28]   Learning Compact Recurrent Neural Networks with Block-Term Tensor Decomposition [J].
Ye, Jinmian ;
Wang, Linnan ;
Li, Guangxi ;
Chen, Di ;
Zhe, Shandian ;
Chu, Xinqi ;
Xu, Zenglin .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9378-9387
[29]   On Compressing Deep Models by Low Rank and Sparse Decomposition [J].
Yu, Xiyu ;
Liu, Tongliang ;
Wang, Xinchao ;
Tao, Dacheng .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :67-76
[30]   Explicit Loss-Error-Aware Quantization for Low-Bit Deep Neural Networks [J].
Zhou, Aojun ;
Yao, Anbang ;
Wang, Kuan ;
Chen, Yurong .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9426-9435