Accelerating Matrix Multiplication in Deep Learning by Using Low-Rank Approximation

被引:8
作者
Osawa, Kazuki [1 ]
Sekiya, Akira [1 ]
Naganuma, Hiroki [1 ]
Yokota, Rio [1 ]
机构
[1] Tokyo Inst Technol, Sch Comp Sci, Tokyo, Japan
来源
2017 INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING & SIMULATION (HPCS) | 2017年
关键词
low-rank approximation; deep learning; image recognition;
D O I
10.1109/HPCS.2017.37
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The open source frameworks of deep learning including TensorFlow, Caffe, Torch, etc. are widely used all over the world and its acceleration have great meaning. In these frameworks, a lot of computation time is spent on convolution, and highly tuned libraries such as cuDNN play important role on accelerating convolution. In these libraries, however, a convolution computation is performed without approximating a dense matrices. In this research, we propose a method to introduce the low-rank approximation method, widely used in the field of scientific and technical computation, into the convolution computation. As a result of investigating the influence on the recognition accuracy of the existing model, it is possible to reduce up to about 90% of rank of data matrices while keeping recognition accuracy 2% of baseline.
引用
收藏
页码:186 / 192
页数:7
相关论文
共 25 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
ANDERSON E., 1999, LAPACK USERSGUIDE, V3rd
[3]  
[Anonymous], 2010, P 13 INT C ART INT S
[4]  
[Anonymous], 2009, Tech. Rep. TR 2009
[5]  
[Anonymous], LECT NOTES COMPUTER
[6]  
[Anonymous], BIGLEARN NIPS WORKSH
[7]  
[Anonymous], 2016, ABS160202830 CORR
[8]  
[Anonymous], 2015, Deep Residual Learning for Image Recognition
[9]  
[Anonymous], 2014, CORR
[10]  
[Anonymous], 2006, 10 INT WORKSHOP FRON