Recursive Binary Neural Network Training Model for Efficient Usage of On-Chip Memory

被引:5
作者
Guan, Tianchan [1 ,2 ]
Liu, Peiye [2 ,3 ]
Zeng, Xiaoyang [1 ]
Kim, Martha [4 ]
Seok, Mingoo [2 ]
机构
[1] Fudan Univ, State Key Lab ASIC & Syst, Shanghai 200433, Peoples R China
[2] Columbia Univ, Dept Elect Engn, New York, NY 10027 USA
[3] Beijing Univ Posts & Telecommun, Beijing Key Lab Intelligent Telecommun Software &, Beijing 100876, Peoples R China
[4] Columbia Univ, Dept Comp Sci, New York, NY 10027 USA
关键词
Deep neural network; binary neural network; deep learning; training acceleration; data storage; IMPLEMENTATION;
D O I
10.1109/TCSI.2019.2895216
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We present a novel deep learning model for a neural network that reduces both computation and data storage overhead. To do so, the proposed model proposes and combines a binary-weight neural network (BNN) training, a storage reuse technique, and an incremental training scheme. The storage requirements can be tuned to meet the desired classification accuracy, storing more parameters on an on-chip memory, and thereby reducing off-chip data storage accesses. Our experiments show 4-6x reduction in weight storage footprint when training binary deep neural network models. On the FPGA platform, this results in a reduced amount of off-chip accesses, enabling our model to train a neural network in 14x shorter latency, as compared to the conventional BNN training method.
引用
收藏
页码:2593 / 2605
页数:13
相关论文
共 53 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   A Circuit-Based Learning Architecture for Multilayer Neural Networks With Memristor Bridge Synapses [J].
Adhikari, Shyam Prasad ;
Kim, Hyongsuk ;
Budhathoki, Ram Kaji ;
Yang, Changju ;
Chua, Leon O. .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2015, 62 (01) :215-223
[3]   YodaNN1 : An Ultra-Low Power Convolutional Neural Network Accelerator Based on Binary Weights [J].
Andri, Renzo ;
Cavigelli, Lukas ;
Rossi, Davide ;
Benini, Luca .
2016 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2016, :236-241
[4]  
[Anonymous], P 3 INT C LEARNING R
[5]  
[Anonymous], 2018, P IEEE INT S CIRC SY
[6]  
[Anonymous], 2015, Nature, DOI [10.1038/nature14539, DOI 10.1038/NATURE14539]
[7]  
[Anonymous], CORR
[8]  
[Anonymous], AM J PSYCHOL
[9]  
[Anonymous], IEEE T PATTERN ANAL
[10]  
[Anonymous], 2016, ARXIV160207360