Knowledge Distillation via Token-Level Relationship Graph Based on the Big Data Technologies

被引:3
|
作者
Zhang, Shuoxi [1 ]
Liu, Hanpeng [1 ]
He, Kun [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, 1037 Luoyu Rd, Wuhan 430074, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge distillation; Graph representation; Graph-based distillation; Big data technology; NEURAL-NETWORKS;
D O I
10.1016/j.bdr.2024.100438
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the big data era, characterized by vast volumes of complex data, the efficiency of machine learning models is of utmost importance, particularly in the context of intelligent agriculture. Knowledge distillation (KD), a technique aimed at both model compression and performance enhancement, serves as a pivotal solution by distilling the knowledge from an elaborate model (teacher) to a lightweight, compact counterpart (student). However, the true potential of KD has not been fully explored. Existing approaches primarily focus on transferring instancelevel information by big data technologies, overlooking the valuable information embedded in token-level relationships, which may be particularly affected by the long-tail effects. To address the above limitations, we propose a novel method called Knowledge Distillation with Token-level Relationship Graph (TRG) that leverages token-wise relationships to enhance the performance of knowledge distillation. By employing TRG, the student model can effectively emulate higher-level semantic information from the teacher model, resulting in improved performance and mobile-friendly efficiency. To further enhance the learning process, we introduce a dynamic temperature adjustment strategy, which encourages the student model to capture the topology structure of the teacher model more effectively. We conduct experiments to evaluate the effectiveness of the proposed method against several state-of-the-art approaches. Empirical results demonstrate the superiority of TRG across various visual tasks, including those involving imbalanced data. Our method consistently outperforms the existing baselines, establishing a new state-of-the-art performance in the field of KD based on big data technologies.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Personalized Federated Learning on long-tailed data via knowledge distillation and generated features
    Lv, Fengling
    Qian, Pinxin
    Lu, Yang
    Wang, Hanzi
    PATTERN RECOGNITION LETTERS, 2024, 186 : 178 - 183
  • [42] Expert-level policy style measurement via knowledge distillation with large language model collaboration
    Zhang, Yujie
    Huang, Biao
    Yuan, Weikang
    Jiang, Zhuoren
    Peng, Longsheng
    Chen, Shuai
    Tan-Soo, Jie-Sheng
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (04)
  • [43] 1-4edGraph-KD: An Effective Federated Graph Learning Scheme Based on Knowledge Distillation
    Wang, Shiyu
    Xie, Jiahao
    Lu, Mingming
    Xiong, Neal N.
    2023 IEEE 9TH INTL CONFERENCE ON BIG DATA SECURITY ON CLOUD, BIGDATASECURITY, IEEE INTL CONFERENCE ON HIGH PERFORMANCE AND SMART COMPUTING, HPSC AND IEEE INTL CONFERENCE ON INTELLIGENT DATA AND SECURITY, IDS, 2023, : 130 - 134
  • [44] SlideGCD: Slide-Based Graph Collaborative Training with Knowledge Distillation for Whole Slide Image Classification
    Shu, Tong
    Shi, Jun
    Sun, Dongdong
    Jiang, Zhiguo
    Zheng, Yushan
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT IV, 2024, 15004 : 470 - 480
  • [45] Relationship-Preserving Knowledge Distillation for Zero-Shot Sketch Based Image Retrieval
    Tian, Jialin
    Xu, Xing
    Wang, Zheng
    Shen, Fumin
    Liu, Xin
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 5473 - 5481
  • [46] KD-INR: Time-Varying Volumetric Data Compression via Knowledge Distillation-Based Implicit Neural Representation
    Han, Jun
    Zheng, Hao
    Bi, Chongke
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (10) : 6826 - 6838
  • [47] Weakly supervised object localization via knowledge distillation based on foreground-background contrast
    Ma, Siteng
    Hou, Biao
    Li, Zhihao
    Wu, Zitong
    Guo, Xianpeng
    Yang, Chen
    Jiao, Licheng
    NEUROCOMPUTING, 2024, 576
  • [48] Knowledge Distillation For CTC-based Speech Recognition Via Consistent Acoustic Representation Learning
    Tian, Sanli
    Deng, Keqi
    Li, Zehan
    Ye, Lingxuan
    Cheng, Gaofeng
    Li, Ta
    Yan, Yonghong
    INTERSPEECH 2022, 2022, : 2633 - 2637
  • [49] An ultra-fast deep-learning-based dose engine for prostate VMAT via knowledge distillation framework with limited patient data
    Tseng, Wenchih
    Liu, Hongcheng
    Yang, Yu
    Liu, Chihray
    Lu, Bo
    PHYSICS IN MEDICINE AND BIOLOGY, 2023, 68 (01)
  • [50] A Multi-Level Adaptive Lightweight Net for Damaged Road Marking Detection Based on Knowledge Distillation
    Wang, Junwei
    Zeng, Xiangqiang
    Wang, Yong
    Ren, Xiang
    Wang, Dongliang
    Qu, Wenqiu
    Liao, Xiaohan
    Pan, Peifen
    REMOTE SENSING, 2024, 16 (14)