Knowledge Distillation via Token-Level Relationship Graph Based on the Big Data Technologies

被引:3
|
作者
Zhang, Shuoxi [1 ]
Liu, Hanpeng [1 ]
He, Kun [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, 1037 Luoyu Rd, Wuhan 430074, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge distillation; Graph representation; Graph-based distillation; Big data technology; NEURAL-NETWORKS;
D O I
10.1016/j.bdr.2024.100438
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the big data era, characterized by vast volumes of complex data, the efficiency of machine learning models is of utmost importance, particularly in the context of intelligent agriculture. Knowledge distillation (KD), a technique aimed at both model compression and performance enhancement, serves as a pivotal solution by distilling the knowledge from an elaborate model (teacher) to a lightweight, compact counterpart (student). However, the true potential of KD has not been fully explored. Existing approaches primarily focus on transferring instancelevel information by big data technologies, overlooking the valuable information embedded in token-level relationships, which may be particularly affected by the long-tail effects. To address the above limitations, we propose a novel method called Knowledge Distillation with Token-level Relationship Graph (TRG) that leverages token-wise relationships to enhance the performance of knowledge distillation. By employing TRG, the student model can effectively emulate higher-level semantic information from the teacher model, resulting in improved performance and mobile-friendly efficiency. To further enhance the learning process, we introduce a dynamic temperature adjustment strategy, which encourages the student model to capture the topology structure of the teacher model more effectively. We conduct experiments to evaluate the effectiveness of the proposed method against several state-of-the-art approaches. Empirical results demonstrate the superiority of TRG across various visual tasks, including those involving imbalanced data. Our method consistently outperforms the existing baselines, establishing a new state-of-the-art performance in the field of KD based on big data technologies.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] FedKG: A Knowledge Distillation-Based Federated Graph Method for Social Bot Detection
    Wang, Xiujuan
    Chen, Kangmiao
    Wang, Keke
    Wang, Zhengxiang
    Zheng, Kangfeng
    Zhang, Jiayue
    SENSORS, 2024, 24 (11)
  • [32] Explicit knowledge transfer of graph-based correlation distillation and diversity data hallucination for few-shot object detection
    Wang, Meng
    Wang, Yang
    Liu, Haipeng
    IMAGE AND VISION COMPUTING, 2024, 143
  • [33] DynamicKD: An effective knowledge distillation via dynamic entropy correction-based distillation for gap optimizing
    Zhu, Songling
    Shang, Ronghua
    Yuan, Bo
    Zhang, Weitong
    Li, Wenjie
    Li, Yangyang
    Jiao, Licheng
    PATTERN RECOGNITION, 2024, 153
  • [34] Online Continual Learning via the Meta-learning update with Multi-scale Knowledge Distillation and Data Augmentation
    Han, Ya-nan
    Liu, Jian-wei
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 113
  • [35] SMIGNN: social recommendation with multi-intention knowledge distillation based on graph neural network
    Yong Niu
    Xing Xing
    Zhichun Jia
    Mindong Xin
    Junye Xing
    The Journal of Supercomputing, 2024, 80 : 6965 - 6988
  • [36] Heterogeneous Defect Prediction Based on Federated Transfer Learning via Knowledge Distillation
    Wang, Aili
    Zhang, Yutong
    Yan, Yixin
    IEEE ACCESS, 2021, 9 : 29530 - 29540
  • [37] SMIGNN: social recommendation with multi-intention knowledge distillation based on graph neural network
    Niu, Yong
    Xing, Xing
    Jia, Zhichun
    Xin, Mindong
    Xing, Junye
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (05) : 6965 - 6988
  • [38] AdaDFKD: Exploring adaptive inter-sample relationship in data-free knowledge distillation
    Li, Jingru
    Zhou, Sheng
    Li, Liangcheng
    Wang, Haishuai
    Bu, Jiajun
    Yu, Zhi
    NEURAL NETWORKS, 2024, 177
  • [39] Semantic segmentation method for continuous images based on multi-level knowledge distillation
    Ling Z.
    Li X.
    Zhang T.
    Chen L.
    Sun L.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2023, 29 (04): : 1244 - 1253
  • [40] Knowledge Distillation based Online Learning Methodology using Unlabeled Data Stream
    Seo, Sanghyun
    Park, Seongchul
    Jeong, Changhoon
    Kim, Juntae
    PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND MACHINE INTELLIGENCE (MLMI 2018), 2018, : 68 - 71