A Dual Rank-Constrained Filter Pruning Approach for Convolutional Neural Networks

被引:12
作者
Fan, Fugui [1 ]
Su, Yuting [1 ]
Jing, Peiguang [1 ]
Lu, Wei [1 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
关键词
Manifolds; Adaptation models; Correlation; Adaptive systems; Adaptive filters; Computer architecture; Information filters; Filter pruning; Grassmann manifold; adaptive affinity graph; low-rank; high-rank; CLASSIFICATION;
D O I
10.1109/LSP.2021.3101670
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Filter pruning has attracted increasing attentions to compress and accelerate the convolutional neural networks (CNNs) on computationally restricted devices. Existing related methods mainly focus on independently leveraging the spatial information of individual filters while ignoring the inner correlation among filters. In this letter, we propose a dual rank-constrained filter pruning approach for convolutional neural networks, in which the representation, clustering, and identification of representative filters are integrated into an adaptive graph regularization framework. Particularly, the proposed approach utilizes the low-rank constraint to capture the low-dimensional intrinsic representations of filters for the adaptive affinity graph construction and clustering. It is noteworthy that the original filters are projected as points on Grassmann manifold for geometrical structure preserving. Meanwhile, the high-rank constraint is employed to select the most informative filter for representing the cluster. Experimental results on CIFAR-10 dataset show that the proposed approach achieves competitive results compared with the state-of-the-art methods by 87.1% in VGGNet-16, 52.4% in GoogLeNet and 72.9% in ResNet-56 in terms of model parameters compression.
引用
收藏
页码:1734 / 1738
页数:5
相关论文
共 26 条
[1]  
Absil PA, 2008, OPTIMIZATION ALGORITHMS ON MATRIX MANIFOLDS, P1
[2]   Automatic Music Stretching Resistance Classification Using Audio Features and Genres [J].
Chen, Jun ;
Wang, Chaokun .
IEEE SIGNAL PROCESSING LETTERS, 2013, 20 (12) :1249-1252
[3]   Approximate statistical tests for comparing supervised classification learning algorithms [J].
Dietterich, TG .
NEURAL COMPUTATION, 1998, 10 (07) :1895-1923
[4]  
Guo XJ, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P3547
[5]   Dictionary Learning and Sparse Coding on Grassmann Manifolds: An Extrinsic Solution [J].
Harandi, Mehrtash ;
Sanderson, Conrad ;
Shen, Chunhua ;
Lovell, Brian C. .
2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, :3120-3127
[6]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[7]   Channel Pruning for Accelerating Very Deep Neural Networks [J].
He, Yihui ;
Zhang, Xiangyu ;
Sun, Jian .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1398-1406
[8]   KRYLOV-SUBSPACE METHODS FOR THE SYLVESTER EQUATION [J].
HU, DY ;
REICHEL, L .
LINEAR ALGEBRA AND ITS APPLICATIONS, 1992, 172 :283-313
[9]  
Krizhevsky Alex, 2009, LEARNING MULTIPLE LA
[10]  
Li H., 2017, INT C LEARN REPR ICL