Deep Convolutional Neural Network Compression via Coupled Tensor Decomposition

被引:16
作者
Sun, Weize [1 ]
Chen, Shaowu [1 ]
Huang, Lei [1 ]
So, Hing Cheung [2 ]
Xie, Min [1 ]
机构
[1] Shenzhen Univ, Coll Elect & Informat Engn, Guangdong Key Lab Intelligent Informat Proc, Shenzhen 518060, Guangdong, Peoples R China
[2] City Univ Hong Kong, Dept Elect Engn, Hong Kong 999077, Peoples R China
基金
中国国家自然科学基金;
关键词
Tensors; Matrix decomposition; Optimization; Neural networks; Image coding; Symmetric matrices; Sparse matrices; Deep neural network; network compression; low rank approximation; coupled tensor decomposition;
D O I
10.1109/JSTSP.2020.3038227
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Large neural networks have aroused impressive progress in various real world applications. However, the expensive storage and computational resources requirement for running deep networks make them problematic to be deployed on mobile devices. Recently, matrix and tensor decompositions have been employed for compressing neural networks. In this paper, we develop a simultaneous tensor decomposition technique for network optimization. The shared network structure is first discussed. Sometimes, not only the structure but also the parameters are shared to form a compressed model at the expense of degraded performance. This indicates that the weight tensors between layers within one network contain both identical components and independent components. To utilize this characteristic, two new coupled tensor train decompositions are developed for fully and partly structure sharing cases, and an alternating optimization approach is proposed for low rank tensor computation. Finally, we restore the performance of the neural network model by fine-tuning. The compression ratio of the devised approach can then be calculated. Experimental results are also included to demonstrate the benefits of our algorithm for both applications of image reconstruction and classification, using the well known datasets such as Cifar-10/Cifar-100 and ImageNet and widely used networks such as ResNet. Comparing to the state-of-the-art independent matrix and tensor decomposition based methods, our model can obtain a better network performance under the same compression ratio.
引用
收藏
页码:603 / 616
页数:14
相关论文
共 52 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
[Anonymous], 2015, ACS SYM SER
[3]  
[Anonymous], 2016, Ultimate tensorization: Compressing convolutional and fc layers alike
[4]  
Astrid M, 2017, INT CONF BIG DATA, P115, DOI 10.1109/BIGCOMP.2017.7881725
[5]   Combining segmental semi-Markov models with neural networks for protein secondary structure prediction [J].
Bidargaddi, Niranjan P. ;
Chetty, Madhu ;
Kamruzzaman, Joarder .
NEUROCOMPUTING, 2009, 72 (16-18) :3943-3950
[6]   Approximate Joint Singular Value Decomposition of an Asymmetric Rectangular Matrix Set [J].
Congedo, Marco ;
Phlypo, Ronald ;
Pham, Dinh-Tuan .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2011, 59 (01) :415-424
[7]   An adaptive augmented Lagrangian method for large-scale constrained optimization [J].
Curtis, Frank E. ;
Jiang, Hao ;
Robinson, Daniel P. .
MATHEMATICAL PROGRAMMING, 2015, 152 (1-2) :201-245
[8]   A multilinear singular value decomposition [J].
De Lathauwer, L ;
De Moor, B ;
Vandewalle, J .
SIAM JOURNAL ON MATRIX ANALYSIS AND APPLICATIONS, 2000, 21 (04) :1253-1278
[9]   A fast and elitist multiobjective genetic algorithm: NSGA-II [J].
Deb, K ;
Pratap, A ;
Agarwal, S ;
Meyarivan, T .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) :182-197
[10]  
Denton E, 2014, ADV NEUR IN, V27