DE-MKD: Decoupled Multi-Teacher Knowledge Distillation Based on Entropy

被引:2
作者
Cheng, Xin [1 ]
Zhang, Zhiqiang [2 ]
Weng, Wei [3 ]
Yu, Wenxin [2 ]
Zhou, Jinjia [1 ]
机构
[1] Hosei Univ, Grad Sch Sci & Engn, Tokyo 1848584, Japan
[2] Southwest Univ Sci & Technol, Sch Sci & Technol, Mianyang 621010, Peoples R China
[3] Kanazawa Univ, Inst Liberal Arts & Sci, Kanazawa 9201192, Japan
关键词
multi-teacher knowledge distillation; image classification; entropy; deep learning;
D O I
10.3390/math12111672
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
The complexity of deep neural network models (DNNs) severely limits their application on devices with limited computing and storage resources. Knowledge distillation (KD) is an attractive model compression technology that can effectively alleviate this problem. Multi-teacher knowledge distillation (MKD) aims to leverage the valuable and diverse knowledge distilled by multiple teacher networks to improve the performance of the student network. Existing approaches typically rely on simple methods such as averaging the prediction logits or using sub-optimal weighting strategies to fuse distilled knowledge from multiple teachers. However, employing these techniques cannot fully reflect the importance of teachers and may even mislead student's learning. To address this issue, we propose a novel Decoupled Multi-Teacher Knowledge Distillation based on Entropy (DE-MKD). DE-MKD decouples the vanilla knowledge distillation loss and assigns adaptive weights to each teacher to reflect its importance based on the entropy of their predictions. Furthermore, we extend the proposed approach to distill the intermediate features from multiple powerful but cumbersome teachers to improve the performance of the lightweight student network. Extensive experiments on the publicly available CIFAR-100 image classification benchmark dataset with various teacher-student network pairs demonstrated the effectiveness and flexibility of our approach. For instance, the VGG8|ShuffleNetV2 model trained by DE-MKD reached 75.25%|78.86% top-one accuracy when choosing VGG13|WRN40-2 as the teacher, setting new performance records. In addition, surprisingly, the distilled student model outperformed the teacher in both teacher-student network pairs.
引用
收藏
页数:10
相关论文
共 35 条
[31]   Be Your Own Teacher: Improve the Performance of Convolutional Neural Networks via Self Distillation [J].
Zhang, Linfeng ;
Song, Jiebo ;
Gao, Anni ;
Chen, Jingwei ;
Bao, Chenglong ;
Ma, Kaisheng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :3712-3721
[32]   ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices [J].
Zhang, Xiangyu ;
Zhou, Xinyu ;
Lin, Mengxiao ;
Sun, Ran .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6848-6856
[33]   Decoupled Knowledge Distillation [J].
Zhao, Borui ;
Cui, Quan ;
Song, Renjie ;
Qiu, Yiyu ;
Liang, Jiajun .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :11943-11952
[34]   Highlight Every Step: Knowledge Distillation via Collaborative Teaching [J].
Zhao, Haoran ;
Sun, Xin ;
Dong, Junyu ;
Chen, Changrui ;
Dong, Zihe .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (04) :2070-2081
[35]   Pyramid Scene Parsing Network [J].
Zhao, Hengshuang ;
Shi, Jianping ;
Qi, Xiaojuan ;
Wang, Xiaogang ;
Jia, Jiaya .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6230-6239