DE-MKD: Decoupled Multi-Teacher Knowledge Distillation Based on Entropy

被引:2
作者
Cheng, Xin [1 ]
Zhang, Zhiqiang [2 ]
Weng, Wei [3 ]
Yu, Wenxin [2 ]
Zhou, Jinjia [1 ]
机构
[1] Hosei Univ, Grad Sch Sci & Engn, Tokyo 1848584, Japan
[2] Southwest Univ Sci & Technol, Sch Sci & Technol, Mianyang 621010, Peoples R China
[3] Kanazawa Univ, Inst Liberal Arts & Sci, Kanazawa 9201192, Japan
关键词
multi-teacher knowledge distillation; image classification; entropy; deep learning;
D O I
10.3390/math12111672
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
The complexity of deep neural network models (DNNs) severely limits their application on devices with limited computing and storage resources. Knowledge distillation (KD) is an attractive model compression technology that can effectively alleviate this problem. Multi-teacher knowledge distillation (MKD) aims to leverage the valuable and diverse knowledge distilled by multiple teacher networks to improve the performance of the student network. Existing approaches typically rely on simple methods such as averaging the prediction logits or using sub-optimal weighting strategies to fuse distilled knowledge from multiple teachers. However, employing these techniques cannot fully reflect the importance of teachers and may even mislead student's learning. To address this issue, we propose a novel Decoupled Multi-Teacher Knowledge Distillation based on Entropy (DE-MKD). DE-MKD decouples the vanilla knowledge distillation loss and assigns adaptive weights to each teacher to reflect its importance based on the entropy of their predictions. Furthermore, we extend the proposed approach to distill the intermediate features from multiple powerful but cumbersome teachers to improve the performance of the lightweight student network. Extensive experiments on the publicly available CIFAR-100 image classification benchmark dataset with various teacher-student network pairs demonstrated the effectiveness and flexibility of our approach. For instance, the VGG8|ShuffleNetV2 model trained by DE-MKD reached 75.25%|78.86% top-one accuracy when choosing VGG13|WRN40-2 as the teacher, setting new performance records. In addition, surprisingly, the distilled student model outperformed the teacher in both teacher-student network pairs.
引用
收藏
页数:10
相关论文
共 35 条
  • [11] Kwon K, 2020, INT CONF ACOUST SPEE, P7409, DOI [10.1109/ICASSP40776.2020.9054698, 10.1109/icassp40776.2020.9054698]
  • [12] Self-supervised knowledge distillation for complementary label learning
    Liu, Jiabin
    Li, Biao
    Lei, Minglong
    Shi, Yong
    [J]. NEURAL NETWORKS, 2022, 155 : 318 - 327
  • [13] Adaptive multi-teacher multi-level knowledge distillation
    Liu, Yuang
    Zhang, Wei
    Wang, Jun
    [J]. NEUROCOMPUTING, 2020, 415 : 106 - 113
  • [14] ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design
    Ma, Ningning
    Zhang, Xiangyu
    Zheng, Hai-Tao
    Sun, Jian
    [J]. COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 : 122 - 138
  • [15] Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
    Ren, Shaoqing
    He, Kaiming
    Girshick, Ross
    Sun, Jian
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) : 1137 - 1149
  • [16] Romero A, 2015, Arxiv, DOI arXiv:1412.6550
  • [17] MobileNetV2: Inverted Residuals and Linear Bottlenecks
    Sandler, Mark
    Howard, Andrew
    Zhu, Menglong
    Zhmoginov, Andrey
    Chen, Liang-Chieh
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4510 - 4520
  • [18] Fully Convolutional Networks for Semantic Segmentation
    Shelhamer, Evan
    Long, Jonathan
    Darrell, Trevor
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (04) : 640 - 651
  • [19] Shen CC, 2019, AAAI CONF ARTIF INTE, P3068
  • [20] Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556