HRank: Filter Pruning using High-Rank Feature Map

被引:649
作者
Lin, Mingbao [1 ]
Ji, Rongrong [1 ,5 ]
Wang, Yan [2 ]
Zhang, Yichen [1 ]
Zhang, Baochang [3 ]
Tian, Yonghong [4 ,5 ]
Shao, Ling [6 ]
机构
[1] Xiamen Univ, Sch Informat, Dept Artificial Intelligence, Media Analyt & Comp Lab, Xiamen, Peoples R China
[2] Pinterest, San Francisco, CA USA
[3] Beihang Univ, Beijing, Peoples R China
[4] Peking Univ, Beijing, Peoples R China
[5] Peng Cheng Lab, Shenzhen, Peoples R China
[6] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2020年
关键词
D O I
10.1109/CVPR42600.2020.00160
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural network pruning offers a promising prospect to facilitate deploying deep neural networks on resource-limited devices. However, existing methods are still challenged by the training inefficiency and labor cost in pruning designs, due to missing theoretical guidance of non-salient network components. In this paper, we propose a novel filter pruning method by exploring the High Rank of feature maps (HRank). Our HRank is inspired by the discovery that the average rank of multiple feature maps generated by a single filter is always the same, regardless of the number of image batches CNNs receive. Based on HRank, we develop a method that is mathematically formulated to prune filters with low-rank feature maps. The principle behind our pruning is that low-rank feature maps contain less information, and thus pruned results can be easily reproduced. Besides, we experimentally show that weights with high-rank feature maps contain more important information, such that even when a portion is not updated, very little damage would be done to the model performance. Without introducing any additional constraints, HRank leads to significant improvements over the state-of-the-arts in terms of FLOPs and parameters reduction, with similar accuracies. For example, with ResNet-110, we achieve a 58.2%-FLOPs reduction by removing 59.2% of the parameters, with only a small loss of 0.14% in top-1 accuracy on CIFAR-10. With Res-50, we achieve a 43.8%-FLOPs reduction by removing 36.7% of the parameters, with only a loss of 1.17% in the top-1 accuracy on ImageNet. The codes can be available at https://github.com/1mbxmu/HRank.
引用
收藏
页码:1526 / 1535
页数:10
相关论文
共 37 条
[1]  
[Anonymous], 2009, LEARNING MULTIPLE LA
[2]  
[Anonymous], 2014, Comput. Sci.
[3]  
[Anonymous], 2016, CORR
[4]  
[Anonymous], 2016, ICLR
[5]  
[Anonymous], 2018, ARXIV180602891
[6]   "Learning-Compression" Algorithms for Neural Net Pruning [J].
Carreira-Perpinan, Miguel A. ;
Idelbayev, Yerlan .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8532-8541
[7]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[8]  
Chen WL, 2015, PR MACH LEARN RES, V37, P2285
[9]   Recent advances in efficient computation of deep convolutional neural networks [J].
Cheng, Jian ;
Wang, Pei-song ;
Li, Gang ;
Hu, Qing-hao ;
Lu, Han-qing .
FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2018, 19 (01) :64-77
[10]  
Denil Misha, 2013, NEURAL INFORM PROCES