Deep compression of convolutional neural networks with low-rank approximation

被引:11
作者
Astrid, Marcella [1 ]
Lee, Seung-Ik [1 ,2 ]
机构
[1] Univ Sci & Technol, Dept Comp Software, Daejeon, South Korea
[2] Elect & Telecommun Res Inst, SW Contents Res Lab, Daejeon, South Korea
关键词
convolutional neural network; CP-decomposition; cyber physical system; model compression; singular value decomposition; tensor power method;
D O I
10.4218/etrij.2018-0065
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The application of deep neural networks (DNNs) to connect the world with cyber physical systems (CPSs) has attracted much attention. However, DNNs require a large amount of memory and computational cost, which hinders their use in the relatively low-end smart devices that are widely used in CPSs. In this paper, we aim to determine whether DNNs can be efficiently deployed and operated in low-end smart devices. To do this, we develop a method to reduce the memory requirement of DNNs and increase the inference speed, while maintaining the performance (for example, accuracy) close to the original level. The parameters of DNNs are decomposed using a hybrid of canonical polyadic-singular value decomposition, approximated using a tensor power method, and fine-tuned by performing iterative one-shot hybrid fine-tuning to recover from a decreased accuracy. In this study, we evaluate our method on frequently used networks. We also present results from extensive experiments on the effects of several fine-tuning methods, the importance of iterative fine-tuning, and decomposition techniques. We demonstrate the effectiveness of the proposed method by deploying compressed networks in smartphones.
引用
收藏
页码:421 / 434
页数:14
相关论文
共 50 条
  • [31] Multi-scale low-rank approximation method for image denoising
    Ou, Yang
    Zhang, Bo
    Li, Bailin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (14) : 20357 - 20371
  • [32] STATISTICAL INFERENCE BASED ON ROBUST LOW-RANK DATA MATRIX APPROXIMATION
    Feng, Xingdong
    He, Xuming
    ANNALS OF STATISTICS, 2014, 42 (01) : 190 - 210
  • [33] Multi-scale low-rank approximation method for image denoising
    Yang Ou
    Bo Zhang
    Bailin Li
    Multimedia Tools and Applications, 2022, 81 : 20357 - 20371
  • [34] Practical Sketching Algorithms for Low-Rank Tucker Approximation of Large Tensors
    Dong, Wandi
    Yu, Gaohang
    Qi, Liqun
    Cai, Xiaohao
    JOURNAL OF SCIENTIFIC COMPUTING, 2023, 95 (02)
  • [35] Low-rank matrix approximation using the Lanczos bidiagonalization process with applications
    Simon, HD
    Zha, HY
    SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2000, 21 (06) : 2257 - 2274
  • [36] Low-Rank Matrix Completion Using Graph Neural Network
    Luong Trung Nguyen
    Shim, Byonghyo
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 17 - 21
  • [37] Tensor Network Methods for Hyperparameter Optimization and Compression of Convolutional Neural Networks
    Naumov, A.
    Melnikov, A.
    Perelshtein, M.
    Melnikov, Ar.
    Abronin, V.
    Oksanichenko, F.
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [38] A Filter Rank Based Pruning Method for Convolutional Neural Networks
    Liu, Hao
    Guan, Zhenyu
    Lei, Peng
    2021 IEEE 20TH INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS (TRUSTCOM 2021), 2021, : 1318 - 1322
  • [39] Review of Lightweight Deep Convolutional Neural Networks
    Chen, Fanghui
    Li, Shouliang
    Han, Jiale
    Ren, Fengyuan
    Yang, Zhen
    ARCHIVES OF COMPUTATIONAL METHODS IN ENGINEERING, 2024, 31 (04) : 1915 - 1937
  • [40] Plankton Classification with Deep Convolutional Neural Networks
    Ouyang Py
    Hu Hong
    Shi Zhongzhi
    2016 IEEE INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC), 2016, : 132 - 136