Short floating-point representation for convolutional neural network inference

被引:7
作者
Hyeong-Ju Kang [1 ]
机构
[1] Korea Univ Technol & Educ, Sch Comp Sci & Engn, Cheonan 31253, Chungnam, South Korea
基金
新加坡国家研究基金会;
关键词
deep learning; convolutional neural networks; number representation; neural network accelerator;
D O I
10.1587/elex.15.20180909
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional neural networks (CNNs) are being widely used in computer vision tasks, and there have been many efforts to implement CNNs in ASIC or FPGA for power-hungry environments. Instead of the previous common representation, the fixed-point representation, this letter proposes a short floating-point representation for CNNs. The short floating-point representation is based on the normal floating-point representation, but has much less width and does not have complex cases like Not-a-Number and infinity cases. The short floating-point representation, contrary to the common belief, can produce a low-complexity computation logic because the operands of the multiplier logic can be shortened by the exponent concept of the floating-point representation. The exponent can also reduce the total length to reduce the SRAM area. The experimental results show that the short floating-point representation with 8-bit total width achieves less-than-1-percentage-point degradation without the aid of retraining in the top-5 accuracy on very deep CNNs of up to 152 layers and gives more than a 60% area reduction in the ASIC implementation.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 26 条
[1]  
[Anonymous], 2016, SQUEEZENET ALEXNET L
[2]  
[Anonymous], 2016, THESIS
[3]  
[Anonymous], 2016, CONVOLUTIONAL NEURAL
[4]  
[Anonymous], PROC CVPR IEEE
[5]  
[Anonymous], 2015, REDUCED PRECISION ST
[6]  
[Anonymous], P INT C LEARN REPR
[7]  
[Anonymous], 2017, DEEP CONVOLUTIONAL N
[8]  
[Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
[9]  
[Anonymous], 2015, P 3 INT C LEARN REPR
[10]  
[Anonymous], 2015, FIXED POINT QUANTIZA