Model Compression Using Progressive Channel Pruning

被引:56
作者
Guo, Jinyang [1 ]
Zhang, Weichen [1 ]
Ouyang, Wanli [1 ]
Xu, Dong [1 ]
机构
[1] Univ Sydney, Sch Elect & Informat Engn, Sydney, NSW 2008, Australia
基金
澳大利亚研究理事会;
关键词
Acceleration; Adaptation models; Convolution; Supervised learning; Neural networks; Computational modeling; Model compression; channel pruning; domain adaptation; transfer learning;
D O I
10.1109/TCSVT.2020.2996231
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this work, we propose a simple but effective channel pruning framework called Progressive Channel Pruning (PCP) to accelerate Convolutional Neural Networks (CNNs). In contrast to the existing channel pruning methods that prune the channels only once per layer in a layer-by-layer fashion, our new progressive framework iteratively prunes a small number of channels from several selected layers, which consists of a three-step attempting-selecting-pruning pipeline in each iteration. In the attempting step, we attempt to prune a pre-defined number of channels from one layer by using any existing channel pruning methods and estimate the accuracy drop for this layer based on the labelled samples in the validation set. In the selecting step, based on the estimated accuracy drops for all layers, we propose a greedy strategy to automatically select a set of layers that will lead to less overall accuracy drop after pruning these layers. In the pruning step, we prune a small number of channels from these selected layers. We further extend our PCP framework to prune channels for the deep transfer learning methods like Domain Adversarial Neural Network (DANN), in which we effectively reduce the data distribution mismatch in the channel pruning process by using both labelled samples from the source domain and pseudo-labelled samples from the target domain. Our comprehensive experiments on two benchmark datasets demonstrate that our PCP framework outperforms the existing channel pruning approaches under both supervised learning and transfer learning settings.
引用
收藏
页码:1114 / 1124
页数:11
相关论文
共 43 条
[1]  
[Anonymous], 2013, INTERSPEECH
[2]  
[Anonymous], 2015, PROC ADVNEURAL INF P
[3]  
[Anonymous], 2015, Advances in neural information processing systems
[4]  
[Anonymous], 2017, Hash code archive-drone delivery
[5]   LCNN: Lookup-based Convolutional Neural Network [J].
Bagherinezhad, Hessam ;
Rastegari, Mohammad ;
Farhadi, Ali .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :860-869
[6]  
Bousmalis K, 2016, ADV NEUR IN, V29
[7]   Unsupervised Pixel-Level Domain Adaptation with Generative Adversarial Networks [J].
Bousmalis, Konstantinos ;
Silberman, Nathan ;
Dohan, David ;
Erhan, Dumitru ;
Krishnan, Dilip .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :95-104
[8]  
Chen SY, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2102
[9]  
Denton E, 2014, ADV NEUR IN, V27
[10]  
Ganin Y., 2014, 3 INT C LEARNING REP