Efficient Neural Network Compression Inspired by Compressive Sensing

被引:12
作者
Gao, Wei [1 ,2 ]
Guo, Yang [1 ,2 ]
Ma, Siwei [3 ]
Li, Ge [1 ,2 ]
Kwong, Sam [4 ]
机构
[1] Peking Univ, Sch Elect & Comp Engn, Shenzhen 518055, Peoples R China
[2] Peng Chong Lab, Shenzhen 518066, Peoples R China
[3] Peking Univ, Inst Digital Media, Beijing 100084, Peoples R China
[4] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
关键词
Transforms; Artificial neural networks; Matrix decomposition; Training; Neural networks; Sparse matrices; Redundancy; Compressive sensing (CS); deep neural networks (DNNs); neural network compression (NNC); two-step training; PURSUIT;
D O I
10.1109/TNNLS.2022.3186008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traditional neural network compression (NNC) methods decrease the model size and floating-point operations (FLOPs) in the manner of screening out unimportant weight parameters; however, the intrinsic sparsity characteristics have not been fully exploited. In this article, from the perspective of signal processing and analysis for network parameters, we propose to use a compressive sensing (CS)-based method, namely NNCS, for performance improvements. Our proposed NNCS is inspired by the discovery that sparsity levels of weight parameters in the transform domain are greater than those in the original domain. First, to achieve sparse representations for parameters in the transform domain during training, we incorporate a constrained CS model into loss function. Second, the proposed effective training process consists of two steps, where the first step trains raw weight parameters and induces and reconstructs their sparse representations and the second step trains transform coefficients to improve network performances. Finally, we transform the entire neural network into another new domain-based representation, and a sparser parameter distribution can be obtained to facilitate inference acceleration. Experimental results demonstrate that NNCS can significantly outperform the other existing state-of-the-art methods in terms of parameter reductions and FLOPs. With VGGNet on CIFAR-10, we decrease 94.8% parameters and achieve a 76.8% reduction of FLOPs, with 0.13% drop in Top-1 accuracy. With ResNet-50 on ImageNet, we decrease 75.6% parameters and achieve a 78.9% reduction of FLOPs, with 1.24% drop in Top-1 accuracy.
引用
收藏
页码:1965 / 1979
页数:15
相关论文
共 67 条
[51]   MobileNetV2: Inverted Residuals and Linear Bottlenecks [J].
Sandler, Mark ;
Howard, Andrew ;
Zhu, Menglong ;
Zhmoginov, Andrey ;
Chen, Liang-Chieh .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4510-4520
[52]  
Schulz A., 2009, SER PUBL MAT COL BRA
[53]   Assembly of rhizosphere microbial communities in Artemisia annua: recruitment of plant growth-promoting microorganisms and inter-kingdom interactions between bacteria and fungi [J].
Shi, Yuhua ;
Pan, Yanshuo ;
Xiang, Li ;
Zhu, Zhihui ;
Fu, Wenbo ;
Hao, Guangfei ;
Geng, Zengchao ;
Chen, Shilin ;
Li, Yuzhong ;
Han, Dongfei .
PLANT AND SOIL, 2022, 470 (1-2) :127-139
[54]   Subspace Matching Pursuit for Sparse Unmixing of Hyperspectral Data [J].
Shi, Zhenwei ;
Tang, Wei ;
Duren, Zhana ;
Jiang, Zhiguo .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2014, 52 (06) :3256-3274
[55]  
Siegel J. W., 2020, ARXIV200809661
[56]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[57]   Manifold Regularized Dynamic Network Pruning [J].
Tang, Yehui ;
Wang, Yunhe ;
Xu, Yixing ;
Deng, Yiping ;
Xu, Chao ;
Tao, Dacheng ;
Xu, Chang .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :5016-5026
[59]   Coordinate descent algorithms [J].
Wright, Stephen J. .
MATHEMATICAL PROGRAMMING, 2015, 151 (01) :3-34
[60]   Quantized Convolutional Neural Networks for Mobile Devices [J].
Wu, Jiaxiang ;
Leng, Cong ;
Wang, Yuhang ;
Hu, Qinghao ;
Cheng, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :4820-4828