Delta-DNN: Efficiently Compressing Deep Neural Networks via Exploiting Floats Similarity

被引:7
|
作者
Hu, Zhenbo [1 ]
Zou, Xiangyu [1 ]
Xia, Wen [1 ,2 ]
Jin, Sian
Tao, Dingwen
Liu, Yang [1 ,2 ]
Zhang, Weizhe [1 ,2 ]
Zhang, Zheng [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen, Peoples R China
[2] Cyberspace Secur Res Ctr, Peng Cheng Lab, Shenzhen, Peoples R China
来源
PROCEEDINGS OF THE 49TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2020 | 2020年
基金
美国国家科学基金会;
关键词
Lossy compression; neural network; delta compression;
D O I
10.1145/3404397.3404408
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep neural networks (DNNs) have gained considerable attention in various real-world applications due to the strong performance on representation learning. However, a DNN needs to be trained many epochs for pursuing a higher inference accuracy, which requires storing sequential versions of DNNs and releasing the updated versions to users. As a result, large amounts of storage and network resources are required, significantly hampering DNN utilization on resource-constrained platforms (e.g., IoT, mobile phone). In this paper, we present a novel delta compression framework called Delta-DNN, which can efficiently compress the float-point numbers in DNNs by exploiting the floats similarity existing in DNNs during training. Specifically, (1) we observe the high similarity of float-point numbers between the neighboring versions of a neural network in training; (2) inspired by delta compression technique, we only record the delta (i.e., the differences) between two neighboring versions, instead of storing the full new version for DNNs; (3) we use the error-bounded lossy compression to compress the delta data for a high compression ratio, where the error bound is strictly assessed by an acceptable loss of DNNs' inference accuracy; (4) we evaluate Delta-DNN's performance on two scenarios, including reducing the transmission of releasing DNNs over the network and saving the storage space occupied by multiple versions of DNNs. According to experimental results on six popular DNNs, Delta-DNN achieves the compression ratio 2x-10x higher than state-of-the-art methods, while without sacrificing inference accuracy and changing the neural network structure.
引用
收藏
页数:12
相关论文
共 13 条
  • [1] Smart-DNN: Efficiently Reducing the Memory Requirements of Running Deep Neural Networks on Resource-constrained Platforms
    Hu, Zhenbo
    Zou, Xiangyu
    Xia, Wen
    Zhao, Yuhong
    Zhang, Weizhe
    Wu, Donglei
    2021 IEEE 39TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2021), 2021, : 533 - 541
  • [2] Short-Term Load Forecasting Using Deep Neural Networks (DNN)
    Hossen, Tareq
    Plathottam, Siby Jose
    Angamuthu, Radha Krishnan
    Ranganathan, Prakash
    Salehfar, Hossein
    2017 NORTH AMERICAN POWER SYMPOSIUM (NAPS), 2017,
  • [3] CFD-based deep neural networks (DNN) model for predicting the hydrodynamics of fluidized beds
    Nadda, Mahesh
    Shah, Suresh Kumar
    Roy, Sangram
    Yadav, Ashutosh
    DIGITAL CHEMICAL ENGINEERING, 2023, 8
  • [4] Beyond Classical Ultrasound Contrast via Deep Neural Networks
    Strohm, Hannah
    Rothluebbers, Sven
    Eickel, Klaus
    Guenther, Matthias
    PROCEEDINGS OF THE 2020 IEEE INTERNATIONAL ULTRASONICS SYMPOSIUM (IUS), 2020,
  • [5] Solar Transposition Modeling via Deep Neural Networks With Sky Images
    Pierce, Benjamin G.
    Braid, Jennifer L.
    Stein, Joshua S.
    Augustyn, Jim
    Riley, Daniel
    IEEE JOURNAL OF PHOTOVOLTAICS, 2022, 12 (01): : 145 - 151
  • [6] QD-Compressor: a Quantization-based Delta Compression Framework for Deep Neural Networks
    Zhang, Shuyu
    Wu, Donglei
    Jin, Haoyu
    Zou, Xiangyu
    Xia, Wen
    Huang, Xiaojia
    2021 IEEE 39TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2021), 2021, : 542 - 550
  • [7] Ultrasonic signal classification for composite materials via deep convolutional neural networks
    Zhang, Qirui
    Guo, Canzhi
    Cheng, Guanggui
    Song, Shoupeng
    Ding, Jianning
    NONDESTRUCTIVE TESTING AND EVALUATION, 2024,
  • [8] Fpar: filter pruning via attention and rank enhancement for deep convolutional neural networks acceleration
    Chen, Yanming
    Wu, Gang
    Shuai, Mingrui
    Lou, Shubin
    Zhang, Yiwen
    An, Zhulin
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2973 - 2985
  • [9] FPC: Filter pruning via the contribution of output feature map for deep convolutional neural networks acceleration
    Chen, Yanming
    Wen, Xiang
    Zhang, Yiwen
    He, Qiang
    KNOWLEDGE-BASED SYSTEMS, 2022, 238
  • [10] Quality Assessment of Friction Welding using Image Super-resolution via Deep Convolutional Neural Networks
    Srinivasan, Kathiravan
    Kumaran, S. Senthil
    MATERIALS TODAY-PROCEEDINGS, 2020, 22 : 2266 - 2273