Generic Automated Implementation of Deep Neural Networks on Field Programmable Gate Arrays

被引:0
作者
Tourad, El Hadrami Cheikh [1 ]
Eleuldj, Mohsine [1 ]
机构
[1] Mohammed V Univ Rabat, Ecole Mohammedia Ingenieurs, Rabat, Morocco
来源
6TH INTERNATIONAL CONFERENCE ON SMART CITY APPLICATIONS | 2022年 / 393卷
关键词
HDL; FPGA; DNN; Design flow;
D O I
10.1007/978-3-030-94191-8_80
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Many deep learning tasks, such as image classification, natural language processing, video analysis, and speech recognition, have been accomplished using DNNs (Deep Neural Networks). However, high-performance deep neural networks' success comes with an increase in computational and memory requirements. Field-Programmable Gate Arrays (FPGA) devices are ideal for deploying DNNs, and they have the appropriate qualifications due to their flexibility, power efficiency, and computing performance. However, DNNs are generally deployed on FPGA using a high-level language such as Python then, manually transformed to Hardware Description Language (HDL) and synthesized using a commercial tool. This method is time-consuming and requires HDL skills, which reduces the use of FPGAs. The paper proposes "DNN2FPGA," a generic design flow to implement the DNN models automatically on the FPG, which can overcome the implementation problem. The article reviews many related works and shows the proposed design flow and hardware implementation. Also, it compares our solution and other recent similar tools. We validate the proposed solution using two case study results: A Multi-Layer Perceptron (MLP) used to solve the classical XOR problem and DNN for MNIST dataset classification. Finally, we present the conclusion and future works. This paper presents a new generic design flow of implementing DNN models automatically from the high-level language to FPGA devices, which takes the model in graph presentation as input and automatically generates the FPGA's hardware implementations.
引用
收藏
页码:989 / 1000
页数:12
相关论文
共 17 条
  • [1] Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
  • [2] [Anonymous], 2016, MICRO
  • [3] Fast inference of deep neural networks in FPGAs for particle physics
    Duarte, J.
    Han, S.
    Harris, P.
    Jindariani, S.
    Kreinar, E.
    Kreis, B.
    Ngadiuba, J.
    Pierini, M.
    Rivera, R.
    Tran, N.
    Wu, Z.
    [J]. JOURNAL OF INSTRUMENTATION, 2018, 13
  • [4] Eleuldj M., 2020, 2020 INT C CLOUD COM
  • [5] Gokhale V, 2017, IEEE INT SYMP CIRC S, P2082
  • [6] FP-DNN: An Automated Framework for Mapping Deep Neural Networks onto FPGAs with RTL-HLS Hybrid Templates
    Guan, Yijin
    Liang, Hao
    Xu, Ningyi
    Wang, Wenqiang
    Shi, Shaoshuai
    Chen, Xi
    Sun, Guangyu
    Zhang, Wei
    Cong, Jason
    [J]. 2017 IEEE 25TH ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM 2017), 2017, : 152 - 159
  • [7] Guo K., 2018, A survey of fpga-based neural network accelerator
  • [8] Liu ZQ, 2016, 2016 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (FPT), P61, DOI 10.1109/FPT.2016.7929190
  • [9] Ma YF, 2017, I C FIELD PROG LOGIC
  • [10] Scalable and Modularized RTL Compilation of Convolutional Neural Networks onto FPGA
    Ma, Yufei
    Suda, Naveen
    Cao, Yu
    Seo, Jae-sun
    Vrudhula, Sarma
    [J]. 2016 26TH INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2016,