FPGA-based acceleration for binary neural networks in edge computing

被引:1
|
作者
Zhan J.-Y. [1 ]
Yu A.-T. [1 ]
Jiang W. [1 ]
Yang Y.-J. [1 ]
Xie X.-N. [2 ]
Chang Z.-W. [3 ]
Yang J.-H. [4 ]
机构
[1] School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu
[2] School of Automation, Chengdu University of Information Technology, Chengdu
[3] State Grid Sichuan Electric Power Research Institute, Chengdu
[4] Department of Information Sciences and Technology, George Mason University, Fairfax
基金
中国国家自然科学基金;
关键词
Accelerator; Binarization; Field-programmable gate array (FPGA); Neural networks; Quantification;
D O I
10.1016/j.jnlest.2023.100204
中图分类号
学科分类号
摘要
As a core component in intelligent edge computing, deep neural networks (DNNs) will increasingly play a critically important role in addressing the intelligence-related issues in the industry domain, like smart factories and autonomous driving. Due to the requirement for a large amount of storage space and computing resources, DNNs are unfavorable for resource-constrained edge computing devices, especially for mobile terminals with scarce energy supply. Binarization of DNN has become a promising technology to achieve a high performance with low resource consumption in edge computing. Field-programmable gate array (FPGA)-based acceleration can further improve the computation efficiency to several times higher compared with the central processing unit (CPU) and graphics processing unit (GPU). This paper gives a brief overview of binary neural networks (BNNs) and the corresponding hardware accelerator designs on edge computing environments, and analyzes some significant studies in detail. The performances of some methods are evaluated through the experiment results, and the latest binarization technologies and hardware acceleration methods are tracked. We first give the background of designing BNNs and present the typical types of BNNs. The FPGA implementation technologies of BNNs are then reviewed. Detailed comparison with experimental evaluation on typical BNNs and their FPGA implementation is further conducted. Finally, certain interesting directions are also illustrated as future work. © 2023 The Authors
引用
收藏
相关论文
共 50 条
  • [31] Energy Efficient FPGA-Based Binary Transformer Accelerator for Edge Devices
    Du, Congpeng
    Ko, Seok-Bum
    Zhang, Hao
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [32] Reliability Evaluation and Analysis of FPGA-Based Neural Network Acceleration System
    Xu, Dawen
    Zhu, Ziyang
    Liu, Cheng
    Wang, Ying
    Zhao, Shuang
    Zhang, Lei
    Liang, Huaguo
    Li, Huawei
    Cheng, Kwang-Ting
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2021, 29 (03) : 472 - 484
  • [33] FPGA-based Acceleration of Time Series Similarity Prediction: From Cloud to Edge
    Kalantar, Amin
    Zimmerman, Zachary
    Brisk, Philip
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2023, 16 (01)
  • [34] Fully Pipelined FPGA Acceleration of Binary Convolutional Neural Networks with Neural Architecture Search
    Ji, Mengfei
    Al-Ars, Zaid
    Chang, Yuchun
    Zhang, Baolin
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2024, 33 (10)
  • [35] A Review of Recent Advances of Binary Neural Networks for Edge Computing
    Zhao W.
    Ma T.
    Gong X.
    Zhang B.
    Doermann D.
    IEEE Journal on Miniaturization for Air and Space Systems, 2021, 2 (01): : 25 - 35
  • [36] Acceleration and implementation of convolutional neural networks based on FPGA
    Zhao, Sijie
    Gao, Shangshang
    Wang, Rugang
    Wang, Yuanyuan
    Zhou, Feng
    Guo, Naihong
    DIGITAL SIGNAL PROCESSING, 2023, 141
  • [37] User Driven FPGA-Based Design Automated Framework of Deep Neural Networks for Low-Power Low-Cost Edge Computing
    Belabed, Tarek
    Coutinho, Maria Gracielly F.
    Fernandes, Marcelo A. C.
    Sakuyama, Carlos Valderrama
    Souani, Chokri
    IEEE ACCESS, 2021, 9 : 89162 - 89180
  • [38] Resource and Data Optimization for Hardware Implementation of Deep Neural Networks Targeting FPGA-based Edge Devices
    Liu, Xinheng
    Kim, Dae Hee
    Wu, Chang
    Chen, Deming
    2018 ACM/IEEE INTERNATIONAL WORKSHOP ON SYSTEM LEVEL INTERCONNECT PREDICTION (SLIP), 2018,
  • [39] FPGA-based implementation of deep neural network using stochastic computing
    Nobari, Maedeh
    Jahanirad, Hadi
    APPLIED SOFT COMPUTING, 2023, 137
  • [40] Binary classification architecture for Edge Computing based on cognitive services and deep neural networks
    Chancusig, Cristian
    Tumbaco, Sergio
    Alulema, Darwin
    Iribarne, Luis
    Criado, Javier
    PROCEEDINGS OF 2022 14TH INTERNATIONAL CONFERENCE ON MANAGEMENT OF DIGITAL ECOSYSTEMS, MEDES 2022, 2022, : 148 - 155