An Accuracy-Preserving Neural Network Compression via Tucker Decomposition

被引:0
作者
Liu, Can [1 ,2 ]
Xie, Kun [1 ,2 ]
Wen, Jigang [3 ]
Xie, Gaogang [4 ]
Li, Kenli [1 ]
机构
[1] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha 410002, Peoples R China
[2] Minist Educ, Key Lab Fus Comp Supercomp & Artificial Intelligen, Changsha 410002, Peoples R China
[3] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411201, Peoples R China
[4] Chinese Acad Sci, Inst Comp Network Informat Ctr, Beijing 100083, Peoples R China
来源
IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING | 2025年 / 10卷 / 02期
基金
中国国家自然科学基金;
关键词
Tensors; Accuracy; Neural networks; Matrix decomposition; Training; Image coding; Redundancy; model compression; Tucker factorization; RANK;
D O I
10.1109/TSUSC.2024.3425962
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has made remarkable progress across many domains, enabled by the capabilities of over-parameterized neural networks with increasing complexity. However, practical applications often necessitate compact and efficient networks because of device constraints. Among recent low-rank decomposition-based neural network compression techniques, Tucker decomposition has emerged as a promising method which effectively compresses the network while preserving the high-order structure and information of the parameters. Despite its promise, designing an efficient Tucker decomposition approach for compressing neural networks while maintaining accuracy is challenging, due to the complexity of setting ranks across multiple layers and the need for extensive fine-tuning. This paper introduces a novel accuracy-aware network compression problem under Tucker decomposition, which considers both network accuracy and compression performance in terms of parameter size. To address this problem, we propose an efficient alternating optimization algorithm that iteratively solves a network training sub-problem and a Tucker decomposition sub-problem to compress the network with performance assurance. The proper Tucker ranks of multiple layers are selected during network training, enabling efficient compression without extensive fine-tuning. We conduct extensive experiments, implementing image classification on five neural networks using four benchmark datasets. The experimental results demonstrate that, without the need for extensive fine-tuning, our proposed method significantly reduces the model size with minimal loss in accuracy, outperforming baseline methods.
引用
收藏
页码:262 / 273
页数:12
相关论文
共 50 条
[31]   Approximately Invertible Neural Network for Learned Image Compression [J].
Gao, Yanbo ;
Li, Shuai ;
Fu, Meng ;
Lv, Chong ;
Yang, Zhiyuan ;
Cai, Xun ;
Yuan, Hui ;
Ye, Mao .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 :3041-3055
[32]   Robust Feature Extraction via l∞-Norm Based Nonnegative Tucker Decomposition [J].
Chen, Bilian ;
Guan, Jiewen ;
Li, Zhening ;
Zhou, Zhehao .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) :7144-7155
[33]   A topology preserving neural network for nonstationary distributions [J].
Nakajima, T ;
Takizawa, H ;
Kobayashi, H ;
Nakamura, T .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 1999, E82D (07) :1131-1135
[34]   THE-V: Verifiable Privacy-Preserving Neural Network via Trusted Homomorphic Execution [J].
Wei, Yuntao ;
Wang, Xueyan ;
Bian, Song ;
Zhao, Weisheng ;
Jin, Yier .
2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
[35]   BAYESIAN LEARNING FOR NEURAL NETWORK COMPRESSION [J].
Chien, Jen-Tzung ;
Chang, Su-Ting .
2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2020,
[36]   Neural Network LFM Pulse Compression [J].
Akhtar, Jabran .
2023 IEEE RADAR CONFERENCE, RADARCONF23, 2023,
[37]   A Programmable Approach to Neural Network Compression [J].
Joseph, Vinu ;
Gopalakrishnan, Ganesh L. ;
Muralidharan, Saurav ;
Garland, Michael ;
Garg, Animesh .
IEEE MICRO, 2020, 40 (05) :17-25
[38]   ARTIFICIAL NEURAL NETWORK FOR IMAGE COMPRESSION [J].
SICURANZA, GL ;
RAMPONI, G ;
MARSI, S .
ELECTRONICS LETTERS, 1990, 26 (07) :477-479
[39]   Floating Point Accuracy Testing in Deep Neural Network Computations via Hypothesis Testing [J].
Wang, Chaojin ;
Shen, Jian .
2020 27TH ASIA-PACIFIC SOFTWARE ENGINEERING CONFERENCE (APSEC 2020), 2020, :257-265
[40]   NEURAL NETWORK FOR SINGULAR VALUE DECOMPOSITION [J].
CICHOCKI, A .
ELECTRONICS LETTERS, 1992, 28 (08) :784-786