Quantized Deep Neural Networks for Energy Efficient Hardware-based Inference

被引:0
作者
Ding, Ruizhou [1 ]
Liu, Zeye [1 ]
Blanton, R. D. [1 ]
Marculescu, Diana [1 ]
机构
[1] Carnegie Mellon Univ, Dept Elect & Comp Engn, Pittsburgh, PA 15213 USA
来源
2018 23RD ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC) | 2018年
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) have been adopted in many systems because of their higher classification accuracy, with custom hardware implementations great candidates for highspeed, accurate inference. While progress in achieving large scale, highly accurate DNNs has been made, significant energy and area are required due to massive memory accesses and computations. Such demands pose a challenge to any DNN implementation, yet it is more natural to handle in a custom hardware platform. To alleviate the increased demand in storage and energy, quantized DNNs constrain their weights (and activations) from floatingpoint numbers to only a few discrete levels. Therefore, storage is reduced, thereby leading to less memory accesses. In this paper, we provide an overview of different types of quantized DNNs, as well as the training approaches for them. Among the various quantized DNNs, our LightNN (Light Neural Network) approach can reduce both memory accesses and computation energy, by filling the gap between classic, full-precision and binarized DNNs. We provide a detailed comparison between LightNNs, conventional DNNs and Binarized Neural Networks (BNNs), with MNIST and CIFAR-10 datasets. In contrast to other quantized DNNs that trade-off significant amounts of accuracy for lower memory requirements, LightNNs can significantly reduce storage, energy and area while still maintaining a test error similar to a large DNN configuration. Thus, LightNNs provide more options for hardware designers to trade-off accuracy and energy.
引用
收藏
页码:1 / 8
页数:8
相关论文
共 30 条
[1]   YodaNN1 : An Ultra-Low Power Convolutional Neural Network Accelerator Based on Binary Weights [J].
Andri, Renzo ;
Cavigelli, Lukas ;
Rossi, Davide ;
Benini, Luca .
2016 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2016, :236-241
[2]  
[Anonymous], 2016, ARXIV160601981
[3]  
[Anonymous], AISTATS
[4]  
[Anonymous], 2011, NEURAL INFORM PROCES
[5]  
[Anonymous], SYN DES COMP PROD VE
[6]  
[Anonymous], ICML JMLR W CP
[7]  
[Anonymous], 2013, Advances in Neural Information Processing Systems
[8]  
[Anonymous], 2016, ADV NEURAL INFORM PR
[9]  
[Anonymous], CVPR
[10]  
[Anonymous], 2016, NIPS