FPGA-Based High-Performance Data Compression Deep Neural Network Accelerator

被引:3
|
作者
Wang, Hanze [1 ]
Fu, Yingxun [1 ]
Ma, Li [1 ]
机构
[1] North China Univ Technol, Coll Informat Sci, Beijing, Peoples R China
关键词
deep neural networks; compression; transmission; fpga;
D O I
10.1109/BDICN55575.2022.00109
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks play an important role in extracting valuable information from massive amounts of data. But these networks require huge computational and memory overhead, which hinders their use in resource-limited environments, such as mobile or embedded devices. In order to solve this problem, researchers usually reduce the amount of data and the number of memory accesses to reduce the overhead caused by data transmission. In this paper, we design a compressed storage and calculation fusion (CSCF) algorithm for massive input data to compress the input data volume and improve the processing efficiency of terminal equipment. Firstly, we scan and compress the collected data, then classify and store the compressed data according to the location of consecutive zero-valued pixel blocks. In order to adapt to actual development scenarios, we choose FPGA hardware architecture with high flexibility, low energy consumption, and short development cycle as the terminal processor. Therefore, we design a classification calculation unit corresponding to classification compression and storage on the FPGA architecture, and improve the performance of the model by fusing the first-layer convolution calculation of the convolution neural network and the compression storage of the input data. The evaluation results show that, compared with the traditional neural network accelerator for uncompressed transmission, our CSCF-FPGA accelerator achieves a speedup of 3.8-4.8 times on the MNIST data set and 1.8-2.1 times on the CIFAR series data set. Small fluctuations in speedup ratio and hardware resource utilization show that CSCF-FPGA not only achieves good performance, but also brings no additional hardware loss.
引用
收藏
页码:563 / 569
页数:7
相关论文
共 50 条
  • [1] A High-performance FPGA-based Accelerator for Gradient Compression
    Ren, Qingqing
    Zhu, Shuyong
    Meng, Xuying
    Zhang, Yujun
    DCC 2022: 2022 DATA COMPRESSION CONFERENCE (DCC), 2022, : 429 - 438
  • [2] High-Performance FPGA-based Accelerator for Bayesian Neural Networks
    Fan, Hongxiang
    Ferianc, Martin
    Rodrigues, Miguel
    Zhou, Hongyu
    Niu, Xinyu
    Luk, Wayne
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1063 - 1068
  • [3] A High Performance FPGA-Based Sorting Accelerator with a Data Compression Mechanism
    Kobayashi, Ryohei
    Kise, Kenji
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2017, E100D (05) : 1003 - 1015
  • [4] VHDL Generator for A High Performance Convolutional Neural Network FPGA-Based Accelerator
    Hamdan, Muhammad K.
    Rover, Diane T.
    2017 INTERNATIONAL CONFERENCE ON RECONFIGURABLE COMPUTING AND FPGAS (RECONFIG), 2017,
  • [5] FPGA-based hardware accelerator for high-performance data-stream processing
    Lysakov K.F.
    Shadrin M.Y.
    Pattern Recognition and Image Analysis, 2013, 23 (1) : 26 - 34
  • [6] An FPGA-based accelerator for deep neural network with novel reconfigurable architecture
    Jia, Han
    Ren, Daming
    Zou, Xuecheng
    IEICE ELECTRONICS EXPRESS, 2021, 18 (04):
  • [7] A High-Performance FPGA-Based Depthwise Separable Convolution Accelerator
    Huang, Jiye
    Liu, Xin
    Guo, Tongdong
    Zhao, Zhijin
    ELECTRONICS, 2023, 12 (07)
  • [8] A High-Efficiency FPGA-Based Accelerator for Binarized Neural Network
    Guo, Peng
    Ma, Hong
    Chen, Ruizhi
    Wang, Donglin
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2019, 28
  • [9] Implementation of FPGA-based Accelerator for Deep Neural Networks
    Tsai, Tsung-Han
    Ho, Yuan-Chen
    Sheu, Ming-Hwa
    2019 IEEE 22ND INTERNATIONAL SYMPOSIUM ON DESIGN AND DIAGNOSTICS OF ELECTRONIC CIRCUITS & SYSTEMS (DDECS), 2019,
  • [10] A High-Performance FPGA-Based Implementation of the LZSS Compression Algorithm
    Shcherbakov, Ivan
    Weis, Christian
    Wehn, Norbert
    2012 IEEE 26TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS & PHD FORUM (IPDPSW), 2012, : 449 - 453