An Accuracy-Preserving Neural Network Compression via Tucker Decomposition

被引:0
|
作者
Liu, Can [1 ,2 ]
Xie, Kun [1 ,2 ]
Wen, Jigang [3 ]
Xie, Gaogang [4 ]
Li, Kenli [1 ]
机构
[1] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha 410002, Peoples R China
[2] Minist Educ, Key Lab Fus Comp Supercomp & Artificial Intelligen, Changsha 410002, Peoples R China
[3] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411201, Peoples R China
[4] Chinese Acad Sci, Inst Comp Network Informat Ctr, Beijing 100083, Peoples R China
来源
IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING | 2025年 / 10卷 / 02期
基金
中国国家自然科学基金;
关键词
Tensors; Accuracy; Neural networks; Matrix decomposition; Training; Image coding; Redundancy; model compression; Tucker factorization; RANK;
D O I
10.1109/TSUSC.2024.3425962
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has made remarkable progress across many domains, enabled by the capabilities of over-parameterized neural networks with increasing complexity. However, practical applications often necessitate compact and efficient networks because of device constraints. Among recent low-rank decomposition-based neural network compression techniques, Tucker decomposition has emerged as a promising method which effectively compresses the network while preserving the high-order structure and information of the parameters. Despite its promise, designing an efficient Tucker decomposition approach for compressing neural networks while maintaining accuracy is challenging, due to the complexity of setting ranks across multiple layers and the need for extensive fine-tuning. This paper introduces a novel accuracy-aware network compression problem under Tucker decomposition, which considers both network accuracy and compression performance in terms of parameter size. To address this problem, we propose an efficient alternating optimization algorithm that iteratively solves a network training sub-problem and a Tucker decomposition sub-problem to compress the network with performance assurance. The proper Tucker ranks of multiple layers are selected during network training, enabling efficient compression without extensive fine-tuning. We conduct extensive experiments, implementing image classification on five neural networks using four benchmark datasets. The experimental results demonstrate that, without the need for extensive fine-tuning, our proposed method significantly reduces the model size with minimal loss in accuracy, outperforming baseline methods.
引用
收藏
页码:262 / 273
页数:12
相关论文
共 50 条
  • [1] Deep neural network compression by Tucker decomposition with nonlinear response
    Liu, Ye
    Ng, Michael K.
    KNOWLEDGE-BASED SYSTEMS, 2022, 241
  • [2] BATUDE: Budget-Aware Neural Network Compression Based on Tucker Decomposition
    Yin, Miao
    Phan, Huy
    Zang, Xiao
    Liao, Siyu
    Yuan, Bo
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8874 - 8882
  • [3] Accuracy-Preserving Calibration via Statistical Modeling on Probability Simplex
    Esaki, Yasushi
    Nakamura, Akihiro
    Kawano, Keisuke
    Tokuhisa, Ryoko
    Kutsuna, Takuro
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [4] Accuracy-preserving Measurement Collection for Realistic Wireless Simulations
    Al-Bado, Mustafa
    Sengul, Cigdem
    Merz, Ruben
    2012 IEEE SYMPOSIUM ON COMPUTERS AND COMMUNICATIONS (ISCC), 2012, : 625 - 630
  • [5] Infrared Image Monitoring Data Compression of Power Distribution Network via Tensor Tucker Decomposition
    Zhao H.
    Feng J.
    Ma L.
    Dianwang Jishu/Power System Technology, 2021, 45 (04): : 1632 - 1639
  • [6] Fast and Accuracy-Preserving Domain Decomposition Methods for Reduced Fracture Models with Nonconforming Time Grids
    Huynh, Phuoc-Toan
    Cao, Yanzhao
    Hoang, Thi-Thao-Phuong
    JOURNAL OF SCIENTIFIC COMPUTING, 2023, 96 (01)
  • [7] Fast and Accuracy-Preserving Domain Decomposition Methods for Reduced Fracture Models with Nonconforming Time Grids
    Phuoc-Toan Huynh
    Yanzhao Cao
    Thi-Thao-Phuong Hoang
    Journal of Scientific Computing, 2023, 96
  • [8] Nested compression of convolutional neural networks with Tucker-2 decomposition
    Zdunek, Rafal
    Gabor, Mateusz
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [9] Deep Convolutional Neural Network Compression via Coupled Tensor Decomposition
    Sun, Weize
    Chen, Shaowu
    Huang, Lei
    So, Hing Cheung
    Xie, Min
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2021, 15 (03) : 603 - 616
  • [10] Geometry- and Accuracy-Preserving Random Forest Proximities
    Rhodes, Jake S.
    Cutler, Adele
    Moon, Kevin R.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (09) : 10947 - 10959