FPGA-Based High-Performance Data Compression Deep Neural Network Accelerator

被引:3
|
作者
Wang, Hanze [1 ]
Fu, Yingxun [1 ]
Ma, Li [1 ]
机构
[1] North China Univ Technol, Coll Informat Sci, Beijing, Peoples R China
来源
2022 INTERNATIONAL CONFERENCE ON BIG DATA, INFORMATION AND COMPUTER NETWORK (BDICN 2022) | 2022年
关键词
deep neural networks; compression; transmission; fpga;
D O I
10.1109/BDICN55575.2022.00109
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks play an important role in extracting valuable information from massive amounts of data. But these networks require huge computational and memory overhead, which hinders their use in resource-limited environments, such as mobile or embedded devices. In order to solve this problem, researchers usually reduce the amount of data and the number of memory accesses to reduce the overhead caused by data transmission. In this paper, we design a compressed storage and calculation fusion (CSCF) algorithm for massive input data to compress the input data volume and improve the processing efficiency of terminal equipment. Firstly, we scan and compress the collected data, then classify and store the compressed data according to the location of consecutive zero-valued pixel blocks. In order to adapt to actual development scenarios, we choose FPGA hardware architecture with high flexibility, low energy consumption, and short development cycle as the terminal processor. Therefore, we design a classification calculation unit corresponding to classification compression and storage on the FPGA architecture, and improve the performance of the model by fusing the first-layer convolution calculation of the convolution neural network and the compression storage of the input data. The evaluation results show that, compared with the traditional neural network accelerator for uncompressed transmission, our CSCF-FPGA accelerator achieves a speedup of 3.8-4.8 times on the MNIST data set and 1.8-2.1 times on the CIFAR series data set. Small fluctuations in speedup ratio and hardware resource utilization show that CSCF-FPGA not only achieves good performance, but also brings no additional hardware loss.
引用
收藏
页码:563 / 569
页数:7
相关论文
共 50 条
  • [31] A FPGA-based Neural Accelerator for Small IoT Devices
    Hong, Seongmin
    Park, Yongjun
    PROCEEDINGS INTERNATIONAL SOC DESIGN CONFERENCE 2017 (ISOCC 2017), 2017, : 294 - 295
  • [32] Exploration and Generation of Efficient FPGA-based Deep Neural Network Accelerators
    Ali, Nermine
    Philippe, Jean-Marc
    Tain, Benoit
    Coussy, Philippe
    2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021), 2021, : 123 - 128
  • [33] The Shunt: An FPGA-Based Accelerator for Network Intrusion Prevention
    Weaver, Nicholas
    Paxson, Vern
    Gonzalez, Jose M.
    FPGA 2007: FIFTEENTH ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2007, : 199 - 206
  • [34] Design and Development of FPGA-based High-Performance Radar Data Stream Mining System
    Liu, Ying
    Ma, Pengshan
    Cui, Hongyuan
    3RD INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND QUANTITATIVE MANAGEMENT, ITQM 2015, 2015, 55 : 876 - 885
  • [35] An FPGA-Based High-Performance Stateful Packet Processing Method
    Lu, Rui
    Guo, Zhichuan
    MICROMACHINES, 2023, 14 (11)
  • [36] A High-Performance, Pipelined, FPGA-Based Genetic Algorithm Machine
    Barry Shackleford
    Greg Snider
    Richard J. Carter
    Etsuko Okushi
    Mitsuhiro Yasuda
    Katsuhiko Seo
    Hiroto Yasuura
    Genetic Programming and Evolvable Machines, 2001, 2 (1) : 33 - 60
  • [37] High-performance Convolutional Neural Network Accelerator Based on Systolic Arrays and Quantization
    Li, Yufeng
    Lu, Shengli
    Luo, Jihe
    Pang, Wei
    Liu, Hao
    2019 IEEE 4TH INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING (ICSIP 2019), 2019, : 335 - 339
  • [38] Optimizing a FPGA-based Neural Accelerator for Small IoT Devices
    Hong, Seongmin
    Lee, Inho
    Park, Yongjun
    2018 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2018, : 176 - 177
  • [39] Performance-oriented FPGA-based convolution neural network designs
    Kao, Chi-Chou
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (14) : 21019 - 21030
  • [40] Performance-oriented FPGA-based convolution neural network designs
    Chi-Chou Kao
    Multimedia Tools and Applications, 2023, 82 : 21019 - 21030