Design Space Exploration for Layer-parallel Execution of Convolutional Neural Networks on CGRAs

被引:8
|
作者
Heidorn, Christian [1 ]
Hannig, Frank [1 ]
Teich, Jurgen [1 ]
机构
[1] Friedrich Alexander Univ Erlangen Nurnberg FAU, Dept Comp Sci, Hardware Software Codesign, Erlangen, Germany
来源
PROCEEDINGS OF THE 23RD INTERNATIONAL WORKSHOP ON SOFTWARE AND COMPILERS FOR EMBEDDED SYSTEMS (SCOPES 2020) | 2020年
关键词
CNN Accelerators; Design Space Exploration; CGRA;
D O I
10.1145/3378678.3391878
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this work, we systematically explore the design space of throughput, energy, and hardware costs for layer-parallel mappings of Convolutional Neural Networks (CNNs) onto coarse-grained reconfigurable arrays (CGRAs). We derive an analytical model that computes the required resources (processing elements) and buffer memory and thus hardware cost C to sustain a given throughput T as well as the resulting overall energy consumption E for inference. Further, we propose an efficient design space exploration (DSE) to determine the fronts of Pareto-optimal (T,E,C) solutions. This exploration helps to determine the limits of scalability of the presented tiled CGRA accelerator architectures in terms of throughput, the number of parallel layers that can be simultaneously processed, and memory requirements. Finally, we provide an evaluation of energy savings achievable on our architecture in comparison to implementations that execute sequentially a CNN layer-by-layer. In experiments, it is shown that layer-parallel processing is able to reduce energy consumption E by 3.6x, hardware cost C by 1.2x, and increase the achievable throughput T by 6.2x for MobileNet.
引用
收藏
页码:26 / 31
页数:6
相关论文
共 32 条
  • [1] AUGER: A Multi-Objective Design Space Exploration Framework for CGRAs
    Li, Jingyuan
    Hu, Yihan
    Dai, Yuan
    Kuang, Huizhen
    Wang, Lingli
    2023 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE TECHNOLOGY, ICFPT, 2023, : 88 - 95
  • [2] ACCDSE: A Design Space Exploration Framework for Convolutional Neural Network Accelerator
    Li, Zhisheng
    Wang, Lei
    Dou, Qiang
    Tang, Yuxing
    Guo, Shasha
    Zhou, Haifang
    Lu, Wenyuan
    COMPUTER ENGINEERING AND TECHNOLOGY, NCCET 2017, 2018, 600 : 22 - 34
  • [3] CNN2Gate: An Implementation of Convolutional Neural Networks Inference on FPGAs with Automated Design Space Exploration
    Ghaffari, Alireza
    Savaria, Yvon
    ELECTRONICS, 2020, 9 (12) : 1 - 23
  • [4] Graph Neural Networks for High-Level Synthesis Design Space Exploration
    Ferretti, Lorenzo
    Cini, Andrea
    Zacharopoulos, Georgios
    Alippi, Cesare
    Pozzi, Laura
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2023, 28 (02)
  • [5] SYSTEMS DESIGN SPACE EXPLORATION BY SERIAL DATAFLOW PROGRAM EXECUTION
    Casale-Brunet, S.
    Mattavelli, M.
    Alberti, C.
    Janneck, J. W.
    2013 ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS AND COMPUTERS, 2013, : 1805 - 1809
  • [6] Design Space Exploration for Orlando Ultra Low-Power Convolutional Neural Network SoC
    Erdem, Ahmet
    Silvano, Cristina
    Boesch, Thomas
    Ornstein, Andrea
    Singh, Surinder-Pal
    Desoli, Giuseppe
    2018 IEEE 29TH INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS (ASAP), 2018, : 113 - 119
  • [7] SpikeExplorer: Hardware-Oriented Design Space Exploration for Spiking Neural Networks on FPGA
    Padovano, Dario
    Carpegna, Alessio
    Savino, Alessandro
    Di Carlo, Stefano
    ELECTRONICS, 2024, 13 (09)
  • [8] Interactive Parallel Coordinates for Parametric Design Space Exploration
    Mohiuddin, Arefin
    Woodbury, Robert
    CHI'20: EXTENDED ABSTRACTS OF THE 2020 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2020,
  • [9] A Design Space Exploration Framework for Deployment of Resource-Constrained Deep Neural Networks
    Zhang, Yan
    Pan, Lei
    Berkowitz, Phillip
    Lee, Mun Wai
    Riggan, Benjamin
    Bhattacharyya, Shuvra S.
    REAL-TIME IMAGE PROCESSING AND DEEP LEARNING 2024, 2024, 13034
  • [10] POSTER: Design Space Exploration for Performance Optimization of Deep Neural Networks on Shared Memory Accelerators
    Venkataramani, Swagath
    Choi, Jungwook
    Srinivasan, Vijayalakshmi
    Gopalakrishnan, Kailash
    Chang, Leland
    2017 26TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT), 2017, : 146 - 147