Collaborative Knowledge Distillation

被引:1
|
作者
Zhang, Weiwei [1 ]
Guo, Yufeng [1 ]
Wang, Junhuang [1 ]
Zhu, Jianqing [1 ]
Zeng, Huanqiang [1 ]
机构
[1] Huaqiao Univ, Coll Engn, Quanzhou 362021, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Knowledge engineering; Training; Feature extraction; Uncertainty; Correlation; Collaboration; Circuits and systems; Knowledge distillation; teacher-student learning; deep learning;
D O I
10.1109/TCSVT.2024.3377251
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Existing research on knowledge distillation has primarily concentrated on the task of facilitating student networks in acquiring the complete knowledge imparted by teacher networks. However, recent studies have shown that good networks are not suitable for acting as teachers, and there is a positive correlation between distillation performance and teacher prediction uncertainty. To address this finding, this paper thoroughly analyzes in depth the reasons why the teacher network affects the distillation performance, gives full play to the participation of the student network in the process of knowledge distillation, and assists the teacher network in distilling the knowledge that is suitable for their learning. In light of this premise, a novel approach known as Collaborative Knowledge Distillation (CKD) is introduced, which is founded upon the concept of "Tailoring the Teaching to the Individual". Compared with Baseline, this paper's method improves students' accuracy by an average of 3.42% in CIFAR-100 experiments, and by an average of 1.71% compared with the classical Knowledge Distillation (KD) method. The ImageNet experiments conducted revealed a significant improvement of 2.04% in the Top-1 accuracy of the students.
引用
收藏
页码:7601 / 7613
页数:13
相关论文
共 50 条
  • [41] Reciprocal Teacher-Student Learning via Forward and Feedback Knowledge Distillation
    Gou, Jianping
    Chen, Yu
    Yu, Baosheng
    Liu, Jinhua
    Du, Lan
    Wan, Shaohua
    Yi, Zhang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7901 - 7916
  • [42] Defect Detection Method Based on Knowledge Distillation
    Zhou, Qunying
    Wang, Hongyuan
    Tang, Ying
    Wang, Yang
    IEEE ACCESS, 2023, 11 : 35866 - 35873
  • [43] Knowledge Distillation Hashing for Occluded Face Retrieval
    Yang, Yuxiang
    Tian, Xing
    Ng, Wing W. Y.
    Gao, Ying
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 9096 - 9107
  • [44] Weighted Knowledge Based Knowledge Distillation
    Kang S.
    Seo K.
    Transactions of the Korean Institute of Electrical Engineers, 2022, 71 (02): : 431 - 435
  • [45] Analysis of Model Compression Using Knowledge Distillation
    Hong, Yu-Wei
    Leu, Jenq-Shiou
    Faisal, Muhamad
    Prakosa, Setya Widyawan
    IEEE ACCESS, 2022, 10 : 85095 - 85105
  • [46] A Virtual Knowledge Distillation via Conditional GAN
    Kim, Sihwan
    IEEE ACCESS, 2022, 10 : 34766 - 34778
  • [47] Selecting Images With Entropy for Frugal Knowledge Distillation
    Kinnas, Michail
    Violos, John
    Karapiperis, Nikolaos Ioannis
    Kompatsiaris, Ioannis
    IEEE ACCESS, 2025, 13 : 28189 - 28203
  • [48] Knowledge Distillation: A Survey
    Huang Z.-H.
    Yang S.-Z.
    Lin W.
    Ni J.
    Sun S.-L.
    Chen Y.-W.
    Tang Y.
    Jisuanji Xuebao/Chinese Journal of Computers, 2022, 45 (03): : 624 - 653
  • [49] Dual-Level Knowledge Distillation via Knowledge Alignment and Correlation
    Ding, Fei
    Yang, Yin
    Hu, Hongxin
    Krovi, Venkat
    Luo, Feng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 2425 - 2435
  • [50] Collaborative knowledge distillation for incomplete multi-view action prediction
    Kumar, Deepak
    Kumar, Chetan
    Shao, Ming
    IMAGE AND VISION COMPUTING, 2021, 107