Collaborative Knowledge Distillation

被引:1
|
作者
Zhang, Weiwei [1 ]
Guo, Yufeng [1 ]
Wang, Junhuang [1 ]
Zhu, Jianqing [1 ]
Zeng, Huanqiang [1 ]
机构
[1] Huaqiao Univ, Coll Engn, Quanzhou 362021, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Knowledge engineering; Training; Feature extraction; Uncertainty; Correlation; Collaboration; Circuits and systems; Knowledge distillation; teacher-student learning; deep learning;
D O I
10.1109/TCSVT.2024.3377251
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Existing research on knowledge distillation has primarily concentrated on the task of facilitating student networks in acquiring the complete knowledge imparted by teacher networks. However, recent studies have shown that good networks are not suitable for acting as teachers, and there is a positive correlation between distillation performance and teacher prediction uncertainty. To address this finding, this paper thoroughly analyzes in depth the reasons why the teacher network affects the distillation performance, gives full play to the participation of the student network in the process of knowledge distillation, and assists the teacher network in distilling the knowledge that is suitable for their learning. In light of this premise, a novel approach known as Collaborative Knowledge Distillation (CKD) is introduced, which is founded upon the concept of "Tailoring the Teaching to the Individual". Compared with Baseline, this paper's method improves students' accuracy by an average of 3.42% in CIFAR-100 experiments, and by an average of 1.71% compared with the classical Knowledge Distillation (KD) method. The ImageNet experiments conducted revealed a significant improvement of 2.04% in the Top-1 accuracy of the students.
引用
收藏
页码:7601 / 7613
页数:13
相关论文
共 50 条
  • [1] Deep Cross-Layer Collaborative Learning Network for Online Knowledge Distillation
    Su, Tongtong
    Liang, Qiyu
    Zhang, Jinsong
    Yu, Zhaoyang
    Xu, Ziyue
    Wang, Gang
    Liu, Xiaoguang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2075 - 2087
  • [2] Collaborative Global-Local Structure Network With Knowledge Distillation for Imbalanced Data Classification
    Wu, Feiyan
    Liu, Zhunga
    Zhang, Zuowei
    Liu, Jiaxiang
    Wang, Longfei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2450 - 2460
  • [3] Highlight Every Step: Knowledge Distillation via Collaborative Teaching
    Zhao, Haoran
    Sun, Xin
    Dong, Junyu
    Chen, Changrui
    Dong, Zihe
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (04) : 2070 - 2081
  • [4] Collaborative knowledge distillation via filter knowledge transfer
    Gou, Jianping
    Hu, Yue
    Sun, Liyuan
    Wang, Zhi
    Ma, Hongxing
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [5] Heterogeneous Knowledge Distillation for Anomaly Detection
    Wu, Longjiang
    Zhou, Jiali
    IEEE ACCESS, 2024, 12 : 161490 - 161499
  • [6] Knowledge Distillation in Acoustic Scene Classification
    Jung, Jee-Weon
    Heo, Hee-Soo
    Shim, Hye-Jin
    Yu, Ha-Jin
    IEEE ACCESS, 2020, 8 : 166870 - 166879
  • [7] Spot-Adaptive Knowledge Distillation
    Song, Jie
    Chen, Ying
    Ye, Jingwen
    Song, Mingli
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 3359 - 3370
  • [8] Feature fusion-based collaborative learning for knowledge distillation
    Li, Yiting
    Sun, Liyuan
    Gou, Jianping
    Du, Lan
    Ou, Weihua
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2021, 17 (11)
  • [9] Transmission Line Detection Through Auxiliary Feature Registration With Knowledge Distillation
    Wang, Yusen
    Zhou, Wujie
    Qian, Xiaohong
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 9413 - 9425
  • [10] Stereo Confidence Estimation via Locally Adaptive Fusion and Knowledge Distillation
    Kim, Sunok
    Kim, Seungryong
    Min, Dongbo
    Frossard, Pascal
    Sohn, Kwanghoon
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 6372 - 6385