An Efficient FPGA-based Overlay Inference Architecture for Fully Connected DNNs

被引:0
作者
Abdelsalam, Ahmed M. [1 ]
Boulet, Felix [1 ]
Demers, Gabriel [1 ]
Langlois, J. M. Pierre [1 ]
Cheriet, Farida [1 ]
机构
[1] Polytech Montreal, Dept Comp & Software Engn, Montreal, PQ, Canada
来源
2018 INTERNATIONAL CONFERENCE ON RECONFIGURABLE COMPUTING AND FPGAS (RECONFIG) | 2018年
关键词
Deep Learning; FPGAs; Hardware Accelerators; Deep Neural Networks; Quantization;
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) have gained significant popularity in several classification and regression applications. The massive computation and memory requirements of DNNs pose special challenges for FPGA implementation. Moreover, programming FPGAs requires hardware-specific knowledge that many machine-learning researchers do not possess. To make the power and versatility of FPGAs available to a wider DNN user community and to improve DNN design efficiency, we introduce a Single hidden layer Neural Network (SNN) multiplication-free overlay architecture with fully connected DNN-level performance. This FPGA inference overlay can be used for applications that are normally solved with fully connected DNNs. The overlay avoids the time needed to synthesize, place, route and regenerate a new bitstream when the application changes. The SNN overlay inputs and activations are quantized to power-of-two values, which allows utilizing shift units instead of multipliers. Since the overlay is a SNN, we fill the FPGA chip with the maximum possible number of neurons that can work in parallel in the hidden layer. On a ZYNQ-7000 ZC706 FPGA, it is thus possible to implement 2450 neurons in the hidden layer and 30 neurons in the output layer. We evaluate the proposed architecture on typical benchmark datasets and demonstrate higher throughput with respect to the state-of-the-art while achieving the same accuracy.
引用
收藏
页数:6
相关论文
共 25 条
[1]  
Abdelsalam A. M., 2017, IEEE FCCM
[2]  
Alemdar H., 2017, IEEE INT JOINT C NEU
[3]  
[Anonymous], 2015, NATURE
[4]  
[Anonymous], 2014, ADV NEURAL INFORM PR
[5]  
Brant A., 2012, IEEE FCCM
[6]  
Courbariaux M., 2016, 160202830 ARXIV
[7]  
Gudovskiy D. A., 2017, 170602393 ARXIV
[8]  
Gupta S., 2015, INT C MACH LEARN AUG
[9]  
Han S., 2015, 151000149 ARXIV
[10]  
Hornik K., 1991, NEURAL NETWORKS