A Block-Floating-Point Arithmetic Based FPGA Accelerator for Convolutional Neural Networks

被引:0
|
作者
Zhang, Heshan [1 ]
Liu, Zhenyu [2 ]
Zhang, Guanwen [1 ]
Dai, Jiwu [1 ]
Lian, Xiaocong [3 ]
Zhou, Wei [1 ]
Ji, Xiangyang [3 ]
机构
[1] Northwestern Polytech Univ, Sch Elect & Informat, Xian, Peoples R China
[2] Tsinghua Univ, RIIT&TNList, Beijing, Peoples R China
[3] Tsinghua Univ, Dept Automat, Beijing, Peoples R China
来源
2019 7TH IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (IEEE GLOBALSIP) | 2019年
基金
中国国家自然科学基金;
关键词
CNN; FPGA; block-floating-point;
D O I
10.1109/globalsip45357.2019.8969292
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Convolutional neural networks (CNNs) have been widely used in computer vision applications and achieved great success. However, large-scale CNN models usually consume a lot of computing and memory resources, which makes it difficult for them to be deployed on embedded devices. An efficient block-floating-point (BFP) arithmetic is proposed in this paper. compared with 32-bit floating-point arithmetic, the memory and off-chip bandwidth requirements during convolution are reduced by 50% and 72.37%, respectively. Due to the adoption of BFP arithmetic, the complex multiplication and addition operations of floating-point numbers can be replaced by the corresponding operations of fixed-point numbers, which is more efficient on hardware. A CNN model can be deployed on our accelerator with no more than 0.14% top-1 accuracy loss, and there is no need for retraining and fine-tuning. By employing a series of ping-pong memory access schemes, 2-dimensional propagate partial multiply-accumulate (PPMAC) processors, and an optimized memory system, we implemented a CNN accelerator on Xilinx VC709 evaluation board. The accelerator achieves a performance of 665.54 GOP/s and a power efficiency of 89.7 GOP/s/W under a 300 MHz working frequency, which outperforms previous FPGA based accelerators significantly.
引用
收藏
页数:5
相关论文
共 50 条
  • [31] FFConv: An FPGA-based Accelerator for Fast Convolution Layers in Convolutional Neural Networks
    Ahmad, Afzal
    Pasha, Muhammad Adeel
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2020, 19 (02)
  • [32] Round-off error in block-floating-point structures
    Rohani, E
    Fakhraie, SM
    ELECTRONICS LETTERS, 2006, 42 (03) : 186 - 187
  • [33] Scalable FPGA Accelerator for Deep Convolutional Neural Networks with Stochastic Streaming
    Alawad, Mohammed
    Lin, Mingjie
    IEEE TRANSACTIONS ON MULTI-SCALE COMPUTING SYSTEMS, 2018, 4 (04): : 888 - 899
  • [34] A FIFO BASED ACCELERATOR FOR CONVOLUTIONAL NEURAL NETWORKS
    Panchbhaiyye, Vineet
    Ogunfunmi, Tokunbo
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 1758 - 1762
  • [35] Throughput-Optimized FPGA Accelerator for Deep Convolutional Neural Networks
    Liu, Zhiqiang
    Dou, Yong
    Jiang, Jingfei
    Xu, Jinwei
    Li, Shijie
    Zhou, Yongmei
    Xu, Yingnan
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2017, 10 (03)
  • [36] A High Performance FPGA-based Accelerator for Large-Scale Convolutional Neural Networks
    Li, Huimin
    Fan, Xitian
    Jiao, Li
    Cao, Wei
    Zhou, Xuegong
    Wang, Lingli
    2016 26TH INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2016,
  • [37] Optimizing FPGA-based Convolutional Neural Networks Accelerator for Image Super-Resolution
    Chang, Jung-Woo
    Kang, Suk-Ju
    2018 23RD ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2018, : 343 - 348
  • [38] Block-floating-point implementation of recursive computations on a multiple datapath DSP
    Kobayashi, Shiro
    Fettweis, Gerhard P.
    IEEE Workshop on Signal Processing Systems, SiPS: Design and Implementation, 1999, : 221 - 230
  • [39] A development and simulation environment for a floating point operations FPGA based accelerator
    Bera, M
    Danese, G
    De Lotto, I
    Leporati, F
    Spelgatti, A
    EUROMICRO SYMPOSIUM ON DIGITAL SYSTEM DESIGN, PROCEEDINGS, 2003, : 173 - 179
  • [40] Precision Exploration of Floating-Point Arithmetic for Spiking Neural Networks
    Kwak, Myeongjin
    Seo, Hyoju
    Kim, Yongtae
    18TH INTERNATIONAL SOC DESIGN CONFERENCE 2021 (ISOCC 2021), 2021, : 71 - 72