Speeding-up and compression convolutional neural networks by low-rank decomposition without fine-tuning

被引:4
作者
Zhang, Meng [1 ]
Liu, Fei [2 ]
Weng, Dongpeng [2 ]
机构
[1] Southeast Univ, Natl ASIC Res Ctr, Sch Elect Sci & Engn, Nanjing 210096, Peoples R China
[2] Southeast Univ, Sch Microelect, Nanjing 210096, Peoples R China
关键词
Convolutional neural networks; Neural network compression; Low-rank decomposition; Rank selection;
D O I
10.1007/s11554-023-01274-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid development of convolutional neural network (CNN), the accuracy of CNN has been significantly improved, which also brings great challenges to the deployment of mobile terminals or embedded devices with limited resources. Recently, significant achievements have been made in compressing CNN through low-rank decomposition. Unlike existing methods that use the same decomposition form and decomposition strategy with fine-tuning based on singular value decomposition (SVD), our method uses different decomposition forms for different layers, and proposes decomposition strategies without fine-tuning. We present a simple and effective scheme to compress the entire CNN, which is called cosine similarity SVD without fine-tuning. For the AlexNet, our cosine similarity algorithm of rank selection takes 84% of the time to find the rank compared with the bayesian optimization (BayesOpt) algorithm. After we tested various CNNs (AlexNet, VGG-16, VGG-19, and ResNet-50) on different data sets, experimental results show that the weight parameter drop can exceed 50% when the accuracy loss is less than 1% without fine-tuning. The floating point operations (FLOPs) drop is about 20%, and the accuracy loss is less than 1% without fine-tuning.
引用
收藏
页数:12
相关论文
共 35 条
[1]  
[Anonymous], 2016, P 24 ACM INT C MULT, DOI 10.1145/2964284.2967280
[2]  
[Anonymous], 2014, CORR
[3]  
[Anonymous], 2014, P BRIT MACH VIS C, DOI DOI 10.5244/C.28.88
[4]  
[Anonymous], 2012, NEURIPS
[5]  
Astrid M, 2017, INT CONF BIG DATA, P115, DOI 10.1109/BIGCOMP.2017.7881725
[6]   Computationally efficient image deblurring using low rank image approximation and its GPU implementation [J].
Chang, Chih-Hsiang ;
Kehtarnavaz, Nasser .
JOURNAL OF REAL-TIME IMAGE PROCESSING, 2016, 12 (03) :567-573
[7]   Recent advances in efficient computation of deep convolutional neural networks [J].
Cheng, Jian ;
Wang, Pei-song ;
Li, Gang ;
Hu, Qing-hao ;
Lu, Han-qing .
FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2018, 19 (01) :64-77
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Denton E, 2014, ADV NEUR IN, V27
[10]   Most Tensor Problems Are NP-Hard [J].
Hillar, Christopher J. ;
Lim, Lek-Heng .
JOURNAL OF THE ACM, 2013, 60 (06)