Deep Model Compression based on the Training History

被引:8
作者
Basha, S. H. Shabbeer [1 ]
Farazuddin, Mohammad [2 ]
Pulabaigari, Viswanath [2 ]
Dubey, Shiv Ram [3 ]
Mukherjee, Snehasis [4 ]
机构
[1] RV Univ, Sch Comp Sci & Engn, Bangalore 560059, Karnataka, India
[2] Indian Inst Informat Technol Sri City, Chittoor 517646, Andhra Prades, India
[3] Indian Inst Informat Technol Allahabad, Prayagraj 211015, Uttar Pradesh, India
[4] Shiv Nadar Univ, Greater Noida 201314, Uttar Pradesh, India
关键词
Convolutional Neural Networks; Filter pruning; Finetuning; Optimization;
D O I
10.1016/j.neucom.2024.127257
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Convolutional Neural Networks (DCNNs) have shown promising performances in several visual recognition problems which motivated the researchers to propose popular architectures such as LeNet, AlexNet, VGGNet, ResNet, and many more. These architectures come at a cost of high computational complexity and parameter storage. To get rid of storage and computational complexity, deep model compression methods have been evolved. We propose a "History Based Filter Pruning (HBFP)" method that utilizes network training history for filter pruning. Specifically, we prune the redundant filters by observing similar patterns in the filter's l(1)-norms (absolute sum of weights) over the training epochs. We iteratively prune the redundant filters of a CNN in three steps. First, we train the model and select the filter pairs with redundant filters in each pair. Next, we optimize the network to ensure an increased measure of similarity between the filters in a pair. This optimization of the network facilitates us to prune one filter from each pair based on its importance without much information loss. Finally, we retrain the network to regain the performance, which is dropped due to filter pruning. We test our approach on popular architectures such as LeNet-5 on MNIST dataset; VGG-16, ResNet-56, and ResNet-110 on CIFAR-10 dataset, and ResNet-50 on ImageNet. The proposed pruning method outperforms the state-of-the-art in terms of FLOPs reduction (floating-point operations) by 97.98 %, 83.42 %, 78.43 %, 74.95 %, and 75.45 % for LeNet-5, VGG-16, ResNet-56, ResNet-110, and ResNet-50, respectively, while maintaining the less error rate. The source code is available at: https://github.com/shabbeersh/HBFP.
引用
收藏
页数:10
相关论文
共 75 条
[1]  
Abbasi-Asl R, 2020, Arxiv, DOI arXiv:1705.07356
[2]  
Alvarez JM, 2016, ADV NEUR IN, V29
[3]  
[Anonymous], 2009, CIFAR-100 Dataset
[4]  
[Anonymous], 2020, IEEE Transactions on Neural Networks and Learning Systems
[5]   Redundant feature pruning for accelerated inference in deep neural networks [J].
Ayinde, Babajide O. ;
Inanc, Tamer ;
Zurada, Jacek M. .
NEURAL NETWORKS, 2019, 118 :148-158
[6]  
Chen WL, 2015, PR MACH LEARN RES, V37, P2285
[7]   Dynamical Channel Pruning by Conditional Accuracy Change for Deep Neural Networks [J].
Chen, Zhiqiang ;
Xu, Ting-Bing ;
Du, Changde ;
Liu, Cheng-Lin ;
He, Huiguang .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (02) :799-813
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]  
Denil Misha, 2013, Adv. Neural Inf. Process. Syst., V26
[10]  
Denton E, 2014, ADV NEUR IN, V27