FPGA-Based High-Performance Data Compression Deep Neural Network Accelerator

被引:3
|
作者
Wang, Hanze [1 ]
Fu, Yingxun [1 ]
Ma, Li [1 ]
机构
[1] North China Univ Technol, Coll Informat Sci, Beijing, Peoples R China
关键词
deep neural networks; compression; transmission; fpga;
D O I
10.1109/BDICN55575.2022.00109
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks play an important role in extracting valuable information from massive amounts of data. But these networks require huge computational and memory overhead, which hinders their use in resource-limited environments, such as mobile or embedded devices. In order to solve this problem, researchers usually reduce the amount of data and the number of memory accesses to reduce the overhead caused by data transmission. In this paper, we design a compressed storage and calculation fusion (CSCF) algorithm for massive input data to compress the input data volume and improve the processing efficiency of terminal equipment. Firstly, we scan and compress the collected data, then classify and store the compressed data according to the location of consecutive zero-valued pixel blocks. In order to adapt to actual development scenarios, we choose FPGA hardware architecture with high flexibility, low energy consumption, and short development cycle as the terminal processor. Therefore, we design a classification calculation unit corresponding to classification compression and storage on the FPGA architecture, and improve the performance of the model by fusing the first-layer convolution calculation of the convolution neural network and the compression storage of the input data. The evaluation results show that, compared with the traditional neural network accelerator for uncompressed transmission, our CSCF-FPGA accelerator achieves a speedup of 3.8-4.8 times on the MNIST data set and 1.8-2.1 times on the CIFAR series data set. Small fluctuations in speedup ratio and hardware resource utilization show that CSCF-FPGA not only achieves good performance, but also brings no additional hardware loss.
引用
收藏
页码:563 / 569
页数:7
相关论文
共 50 条
  • [11] A reconfigurable FPGA-based spiking neural network accelerator
    Yin, Mingqi
    Cui, Xiaole
    Wei, Feng
    Liu, Hanqing
    Jiang, Yuanyuan
    Cui, Xiaoxin
    MICROELECTRONICS JOURNAL, 2024, 152
  • [12] FPGA-Based High-Performance Network Impairment Emulator
    Duan, Dexuan
    Wang, Xinshuo
    Li, Lin
    Liu, Lei
    ELECTRONICS, 2024, 13 (24):
  • [13] Implementation of Data-optimized FPGA-based Accelerator for Convolutional Neural Network
    Cho, Mannhee
    Kim, Youngmin
    2020 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2020,
  • [14] Deep Neural Network Accelerator based on FPGA
    Thang Viet Huynh
    2017 4TH NAFOSTED CONFERENCE ON INFORMATION AND COMPUTER SCIENCE (NICS), 2017, : 254 - 257
  • [15] An FPGA-Based High-Throughput Dataflow Accelerator for Lightweight Neural Network
    Zhao, Zhiyuan
    Li, Jixing
    Chen, Gang
    Jiang, Zhelong
    Qiao, Ruixiu
    Xu, Peng
    Chen, Yihao
    Lu, Huaxiang
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [16] Fast FPGA-based Emulation for ReRAM-Enabled Deep Neural Network Accelerator
    Shi, Yongquan
    Sun, Yongshuai
    Jiang, Jianfei
    He, Guanghui
    Wang, Qin
    Jing, Naifeng
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [17] An FPGA-based Accelerator Implementation for Deep Convolutional Neural Networks
    Zhou, Yongmei
    Jiang, Jingfei
    PROCEEDINGS OF 2015 4TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND NETWORK TECHNOLOGY (ICCSNT 2015), 2015, : 829 - 832
  • [18] Composite FPGA-based Accelerator for Deep Convolutional Neural Networks
    HuanZhang
    YuanYang
    YangXiao
    2019 IEEE INTERNATIONAL CONFERENCE ON ELECTRON DEVICES AND SOLID-STATE CIRCUITS (EDSSC), 2019,
  • [19] An FPGA-based Accelerator Platform Implements for Convolutional Neural Network
    Meng, Xiao
    Yu, Lixin
    Qin, Zhiyong
    2019 THE 3RD INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPILATION, COMPUTING AND COMMUNICATIONS (HP3C 2019), 2019, : 25 - 28
  • [20] Using Data Compression for Optimizing FPGA-Based Convolutional Neural Network Accelerators
    Guan, Yijin
    Xu, Ningyi
    Zhang, Chen
    Yuan, Zhihang
    Cong, Jason
    ADVANCED PARALLEL PROCESSING TECHNOLOGIES, 2017, 10561 : 14 - 26