Fine-Tuning Channel-Pruned Deep Model via Knowledge Distillation

被引:0
作者
Zhang, Chong [1 ]
Wang, Hong-Zhi [1 ]
Liu, Hong-Wei [1 ]
Chen, Yi-Lin [2 ]
机构
[1] Harbin Inst Technol, Fac Comp, Harbin 150001, Peoples R China
[2] Harbin Inst Technol, Sch Astronaut, Harbin 150001, Peoples R China
基金
中国国家自然科学基金;
关键词
model compression; deep learning; knowledge distillation; fine-tuning;
D O I
10.1007/s11390-023-2386-8
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep convolutional neural networks with high performance are hard to be deployed in many real world applications, since the computing resources of edge devices such as smart phones or embedded GPU are limited. To alleviate this hardware limitation, the compression of deep neural networks from the model side becomes important. As one of the most popular methods in the spotlight, channel pruning of the deep convolutional model can effectively remove redundant convolutional channels from the CNN (convolutional neural network) without affecting the network's performance remarkably. Existing methods focus on pruning design, evaluating the importance of different convolutional filters in the CNN model. A fast and effective fine-tuning method to restore accuracy is urgently needed. In this paper, we propose a fine-tuning method KDFT (Knowledge Distillation Based Fine-Tuning), which improves the accuracy of fine-tuned models with almost negligible training overhead by introducing knowledge distillation. Extensive experimental results on benchmark datasets with representative CNN models show that up to 4.86% accuracy improvement and 79% time saving can be obtained.
引用
收藏
页码:1238 / 1247
页数:10
相关论文
共 36 条
[1]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[2]   Distilling Knowledge via Knowledge Review [J].
Chen, Pengguang ;
Liu, Shu ;
Zhao, Hengshuang ;
Jia, Jiaya .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :5006-5015
[3]  
Chen WL, 2015, PR MACH LEARN RES, V37, P2285
[4]   Towards Efficient Model Compression via Learned Global Ranking [J].
Chin, Ting-Wu ;
Ding, Ruizhou ;
Zhang, Cha ;
Marculescu, Diana .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1515-1525
[5]  
Denil M., 2013, ADV NEURAL INFORM PR, P2148
[6]  
Denton E, 2014, ADV NEUR IN, V27
[7]   Discrete Model Compression with Resource Constraint for Deep Neural Networks [J].
Gao, Shangqian ;
Huang, Feihu ;
Pei, Jian ;
Huang, Heng .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1896-1905
[8]   Artificial neural networks (the multilayer perceptron) - A review of applications in the atmospheric sciences [J].
Gardner, MW ;
Dorling, SR .
ATMOSPHERIC ENVIRONMENT, 1998, 32 (14-15) :2627-2636
[9]   Multi-Dimensional Pruning: A Unified Framework for Model Compression [J].
Guo, Jinyang ;
Ouyang, Wanli ;
Xu, Dong .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1505-1514
[10]  
Guo YW, 2016, ADV NEUR IN, V29