A Feature Map Lossless Compression Framework for Convolutional Neural Network Accelerators

被引:0
|
作者
Zhang, Zekun [1 ,2 ]
Jiao, Xin [2 ]
Xu, Chengyu [2 ]
机构
[1] Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
[2] SenseTime Res, Shanghai, Peoples R China
关键词
Feature map compression; deep learning; convolutional neural networks; hardware acceleration;
D O I
10.1109/AICAS59952.2024.10595980
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a predictor-based lossless compression algorithm for the feature maps present within convolutional neural networks (CNNs), which provides the possibility to solve the system bandwidth bottleneck and excessive power consumption problem of hardware acceleration. It is also an algorithm-hardware co-design methodology, yielding a hardware-friendly compression approach with low power consumption. The performance of the algorithm is evaluated in the detection, recognition, and segment CNN tasks respectively. Results show that an average compression ratio of 3.03x and a gain of nearly 50% over existing methods can be achieved for VGG-16; 2.78x and a gain of around 51% for ResNet-18; 2.45 and a gain of nearly 38% for SegNet.
引用
收藏
页码:422 / 426
页数:5
相关论文
共 50 条
  • [1] Sparse convolutional neural network acceleration with lossless input feature map compression for resource-constrained systems
    Kwon, Jisu
    Kong, Joonho
    Munir, Arslan
    IET COMPUTERS AND DIGITAL TECHNIQUES, 2022, 16 (01): : 29 - 43
  • [2] Area Efficient Compression for Floating-Point Feature Maps in Convolutional Neural Network Accelerators
    Yan, Bai-Kui
    Ruan, Shanq-Jang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (02) : 746 - 750
  • [3] Extended Bit-Plane Compression for Convolutional Neural Network Accelerators
    Cavigelli, Lukas
    Benini, Luca
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 279 - 283
  • [4] An efficient loop tiling framework for convolutional neural network inference accelerators
    Huang, Hongmin
    Hu, Xianghong
    Li, Xueming
    Xiong, Xiaoming
    IET CIRCUITS DEVICES & SYSTEMS, 2022, 16 (01) : 116 - 123
  • [5] Convolutional neural network simplification via feature map pruning
    Zou, Junhua
    Rui, Ting
    Zhou, You
    Yang, Chengsong
    Zhang, Sai
    COMPUTERS & ELECTRICAL ENGINEERING, 2018, 70 : 950 - 958
  • [6] A Lightweight Convolutional Neural Network Architecture with Slice Feature Map
    Zhang Y.
    Zheng Z.
    Liu H.
    Xiang D.
    He X.
    Li Z.
    He Y.
    Khodja A.E.
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2019, 32 (03): : 237 - 246
  • [7] Structured feature sparsity training for convolutional neural network compression
    Wang, Wei
    Zhu, Liqiang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2020, 71
  • [8] Synapse Compression for Event-Based Convolutional-Neural-Network Accelerators
    Bamberg, Lennart
    Pourtaherian, Arash
    Waeijen, Luc
    Chahar, Anupam
    Moreira, Orlando
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (04) : 1227 - 1240
  • [9] Using Data Compression for Optimizing FPGA-Based Convolutional Neural Network Accelerators
    Guan, Yijin
    Xu, Ningyi
    Zhang, Chen
    Yuan, Zhihang
    Cong, Jason
    ADVANCED PARALLEL PROCESSING TECHNOLOGIES, 2017, 10561 : 14 - 26
  • [10] ASC: Adaptive Scale Feature Map Compression for Deep Neural Network
    Yao, Yuan
    Chang, Tian-Sheuan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (03) : 1417 - 1428