Deep compression of convolutional neural networks with low-rank approximation

被引:11
作者
Astrid, Marcella [1 ]
Lee, Seung-Ik [1 ,2 ]
机构
[1] Univ Sci & Technol, Dept Comp Software, Daejeon, South Korea
[2] Elect & Telecommun Res Inst, SW Contents Res Lab, Daejeon, South Korea
关键词
convolutional neural network; CP-decomposition; cyber physical system; model compression; singular value decomposition; tensor power method;
D O I
10.4218/etrij.2018-0065
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The application of deep neural networks (DNNs) to connect the world with cyber physical systems (CPSs) has attracted much attention. However, DNNs require a large amount of memory and computational cost, which hinders their use in the relatively low-end smart devices that are widely used in CPSs. In this paper, we aim to determine whether DNNs can be efficiently deployed and operated in low-end smart devices. To do this, we develop a method to reduce the memory requirement of DNNs and increase the inference speed, while maintaining the performance (for example, accuracy) close to the original level. The parameters of DNNs are decomposed using a hybrid of canonical polyadic-singular value decomposition, approximated using a tensor power method, and fine-tuned by performing iterative one-shot hybrid fine-tuning to recover from a decreased accuracy. In this study, we evaluate our method on frequently used networks. We also present results from extensive experiments on the effects of several fine-tuning methods, the importance of iterative fine-tuning, and decomposition techniques. We demonstrate the effectiveness of the proposed method by deploying compressed networks in smartphones.
引用
收藏
页码:421 / 434
页数:14
相关论文
共 50 条
  • [41] Deep Compression with Low Rank and Sparse Integrated Decomposition
    Huang, Junhao
    Sun, Weize
    Huang, Lei
    Chen, Shaowu
    PROCEEDINGS OF 2019 IEEE 7TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND NETWORK TECHNOLOGY (ICCSNT 2019), 2019, : 289 - 292
  • [42] Structured Compression of Convolutional Neural Networks for Specialized Tasks
    Gabbay, Freddy
    Salomon, Benjamin
    Shomron, Gil
    MATHEMATICS, 2022, 10 (19)
  • [43] Tensor total variation regularised low-rank approximation framework for video deraining
    Baiju, P. S.
    Jayan, P. Deepak
    George, Sudhish N.
    IET IMAGE PROCESSING, 2020, 14 (14) : 3602 - 3612
  • [44] Similarity Measure based on Low-Rank Approximation for Highly Scalable Recommender Systems
    Seifzadeh, Sepideh
    Miri, Ali
    2015 IEEE TRUSTCOM/BIGDATASE/ISPA, VOL 2, 2015, : 66 - 71
  • [45] Mixed precision low-rank approximations and their application to block low-rank LU factorization
    Amestoy, Patrick
    Boiteau, Olivier
    Buttari, Alfredo
    Gerest, Matthieu
    Jezequel, Fabienne
    L'excellent, Jean-Yves
    Mary, Theo
    IMA JOURNAL OF NUMERICAL ANALYSIS, 2023, 43 (04) : 2198 - 2227
  • [46] A survey of model compression for deep neural networks
    Li J.-Y.
    Zhao Y.-K.
    Xue Z.-E.
    Cai Z.
    Li Q.
    Gongcheng Kexue Xuebao/Chinese Journal of Engineering, 2019, 41 (10): : 1229 - 1239
  • [47] Deep Compression on Convolutional Neural Network for Artistic Style Transfer
    Hu, Jian
    He, Kun
    Hopcroft, John E.
    Zhang, Yaren
    THEORETICAL COMPUTER SCIENCE, NCTCS 2017, 2017, 768 : 157 - 166
  • [48] Model Compression for Deep Neural Networks: A Survey
    Li, Zhuo
    Li, Hengyi
    Meng, Lin
    COMPUTERS, 2023, 12 (03)
  • [49] The impossibility of low-rank representations for triangle-rich complex networks
    Seshadhri, C.
    Sharma, Aneesh
    Stolman, Andrew
    Goel, Ashish
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2020, 117 (11) : 5631 - 5637
  • [50] Automatic Compression Ratio Allocation for Pruning Convolutional Neural Networks
    Liu, Yunfeng
    Kong, Huihui
    Yu, Peihua
    ICVISP 2019: PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING, 2019,