Exploration and Generation of Efficient FPGA-based Deep Neural Network Accelerators

被引:4
|
作者
Ali, Nermine [1 ]
Philippe, Jean-Marc [1 ]
Tain, Benoit [1 ]
Coussy, Philippe [2 ]
机构
[1] Univ Paris Saclay, CEA, List, F-91120 Palaiseau, France
[2] Univ South Brittany, Lorient, France
关键词
Convolutional Neural Networks; Design Space Exploration; High Level Synthesis; Embedded Systems; FPGA;
D O I
10.1109/SiPS52927.2021.00030
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional Neural Networks (CNNs) have emerged as an answer to next-generation applications such as complex image recognition and object detection. Embedding such compute-intensive and memory-hungry algorithms on edge systems will lead to smarter high-value applications. However, the algorithmic innovations in the CNN field leave the hardware accelerators one step behind. Reconfigurable hardware (e.g. FPGAs) allows designing custom accelerators adapted to new algorithms. Furthermore, new design approaches such as high-level synthesis (HLS) enable to generate RTL code based on high-level function descriptions. This paper presents a high-level CNN accelerator generation framework for FPGAs. A first phase of the framework characterizes CNN descriptions using hardware-aware metrics. These metrics then drive a hardware generation phase which builds the proper C source code implementation for each layer of the network. Finally, an HLS tool outputs the synthesizable RTL code of the accelerator. This approach aims at reducing the gap between the evolving applications based on artificial intelligence and hardware accelerators, thus reducing time-to-market of new systems.
引用
收藏
页码:123 / 128
页数:6
相关论文
共 50 条
  • [1] DeepBurning: Automatic Generation of FPGA-based Learning Accelerators for the Neural Network Family
    Wang, Ying
    Xu, Jie
    Han, Yinhe
    Li, Huawei
    Li, Xiaowei
    2016 ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2016,
  • [2] FPGA-Based Reduction Techniques for Efficient Deep Neural Network Deployment
    Page, Adam
    Mohsenin, Tinoosh
    2016 IEEE 24TH ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM), 2016, : 200 - 200
  • [3] [DL] A Survey of FPGA-based Neural Network Inference Accelerators
    Guo, Kaiyuan
    Zeng, Shulin
    Yu, Jincheng
    Wang, Yu
    Yang, Huazhong
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2019, 12 (01)
  • [4] Efficient Deep Neural Network Acceleration through FPGA-based Batch Processing
    Posewsky, Thorbjoern
    Ziener, Daniel
    2016 INTERNATIONAL CONFERENCE ON RECONFIGURABLE COMPUTING AND FPGAS (RECONFIG16), 2016,
  • [5] Design Space Exploration of HW Accelerators and Network Infrastructure for FPGA-Based MPSoC
    Dammak, Bouthaina
    Baklouti, Mouna
    Alsekait, Deema
    IEEE ACCESS, 2024, 12 : 15280 - 15289
  • [6] Towards Efficient Design Space Exploration of FPGA-based Accelerators for Streaming HPC Applications
    Koraei, Mostafa
    Jahre, Magnus
    Fatemi, S. Omid
    FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, : 287 - 287
  • [7] Using Data Compression for Optimizing FPGA-Based Convolutional Neural Network Accelerators
    Guan, Yijin
    Xu, Ningyi
    Zhang, Chen
    Yuan, Zhihang
    Cong, Jason
    ADVANCED PARALLEL PROCESSING TECHNOLOGIES, 2017, 10561 : 14 - 26
  • [8] Design Space Exploration of FPGA-Based Deep Convolutional Neural Networks
    Motamedi, Mohammad
    Gysel, Philipp
    Akella, Venkatesh
    Ghiasi, Soheil
    2016 21ST ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2016, : 575 - 580
  • [9] The Progress and Trends of FPGA-Based Accelerators in Deep Learning
    Wu Y.-X.
    Liang K.
    Liu Y.
    Cui H.-M.
    Jisuanji Xuebao/Chinese Journal of Computers, 2019, 42 (11): : 2461 - 2480
  • [10] Throughput optimizations for FPGA-based deep neural network inference
    Posewsky, Thorbjoern
    Ziener, Daniel
    MICROPROCESSORS AND MICROSYSTEMS, 2018, 60 : 151 - 161