Learning Filter Pruning Criteria for Deep Convolutional Neural Networks Acceleration

被引:198
作者
He, Yang [1 ]
Ding, Yuhang [2 ]
Liu, Ping [1 ]
Zhu, Linchao [1 ]
Zhang, Hanwang [3 ]
Yang, Yi [1 ]
机构
[1] Univ Technol Sydney, ReLER, Sydney, NSW, Australia
[2] Baidu Res, Beijing, Peoples R China
[3] Nanyang Technol Univ, Singapore, Singapore
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2020年
关键词
D O I
10.1109/CVPR42600.2020.00208
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Filter pruning has been widely applied to neural network compression and acceleration. Existing methods usually utilize pre-defined pruning criteria, such as l(p)-norm, to prune unimportant filters. There are two major limitations to these methods. First, prevailing methods fail to consider the variety of filter distribution across layers. To extract features of the coarse level to the fine level, the filters of different layers have various distributions. Therefore, it is not suitable to utilize the same pruning criteria to different functional layers. Second, prevailing layer-by-layer pruning methods process each layer independently and sequentially, failing to consider that all the layers in the network collaboratively make the final prediction. In this paper, we propose Learning Filter Pruning Criteria (LFPC) to solve the above problems. Specifically, we develop a differentiable pruning criteria sampler. This sampler is learnable and optimized by the validation loss of the pruned network obtained from the sampled criteria. In this way, we could adaptively select the appropriate pruning criteria for different functional layers. Besides, when evaluating the sampled criteria, LFPC comprehensively considers the contribution of all the layers at the same time. Experiments validate our approach on three image classification benchmarks. Notably, on ILSVRC-2012, our LFPC reduces more than 60% FLOPs on ResNet-50 with only 0.83% top-5 accuracy loss.
引用
收藏
页码:2006 / 2015
页数:10
相关论文
共 58 条
[11]  
Choi HS, 2017, ASIA-PAC INT SYM ELE, P4, DOI 10.1109/APEMC.2017.7975409
[12]   AutoAugment: Learning Augmentation Strategies from Data [J].
Cubuk, Ekin D. ;
Zoph, Barret ;
Mane, Dandelion ;
Vasudevan, Vijay ;
Le, Quoc V. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :113-123
[13]   Centripetal SGD for Pruning Very Deep Convolutional Networks With Complicated Structure [J].
Ding, Xiaohan ;
Ding, Guiguang ;
Guo, Yuchen ;
Han, Jungong .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :4938-4948
[14]  
Dong X., 2017, CVPR
[15]  
Dong X, 2017, ADV NEUR IN, V30
[16]  
Dong XY, 2019, ADV NEUR IN, V32
[17]   Searching for A Robust Neural Architecture in Four GPU Hours [J].
Dong, Xuanyi ;
Yang, Yi .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1761-1770
[18]  
Dubey A., 2018, ARXIV180709810
[19]  
Frankle J., 2019, ICLR
[20]  
Guo Y., 2019, INT C MACHINE LEARNI, P1607