Knowledge Distillation via Token-Level Relationship Graph Based on the Big Data Technologies

被引:3
|
作者
Zhang, Shuoxi [1 ]
Liu, Hanpeng [1 ]
He, Kun [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, 1037 Luoyu Rd, Wuhan 430074, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge distillation; Graph representation; Graph-based distillation; Big data technology; NEURAL-NETWORKS;
D O I
10.1016/j.bdr.2024.100438
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the big data era, characterized by vast volumes of complex data, the efficiency of machine learning models is of utmost importance, particularly in the context of intelligent agriculture. Knowledge distillation (KD), a technique aimed at both model compression and performance enhancement, serves as a pivotal solution by distilling the knowledge from an elaborate model (teacher) to a lightweight, compact counterpart (student). However, the true potential of KD has not been fully explored. Existing approaches primarily focus on transferring instancelevel information by big data technologies, overlooking the valuable information embedded in token-level relationships, which may be particularly affected by the long-tail effects. To address the above limitations, we propose a novel method called Knowledge Distillation with Token-level Relationship Graph (TRG) that leverages token-wise relationships to enhance the performance of knowledge distillation. By employing TRG, the student model can effectively emulate higher-level semantic information from the teacher model, resulting in improved performance and mobile-friendly efficiency. To further enhance the learning process, we introduce a dynamic temperature adjustment strategy, which encourages the student model to capture the topology structure of the teacher model more effectively. We conduct experiments to evaluate the effectiveness of the proposed method against several state-of-the-art approaches. Empirical results demonstrate the superiority of TRG across various visual tasks, including those involving imbalanced data. Our method consistently outperforms the existing baselines, establishing a new state-of-the-art performance in the field of KD based on big data technologies.
引用
收藏
页数:12
相关论文
共 50 条
  • [21] KDCRec: Knowledge Distillation for Counterfactual Recommendation via Uniform Data
    Liu, Dugang
    Cheng, Pengxiang
    Lin, Zinan
    Luo, Jinwei
    Dong, Zhenhua
    He, Xiuqiang
    Pan, Weike
    Ming, Zhong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (08) : 8143 - 8156
  • [22] A two-phase knowledge distillation model for graph convolutional network-based recommendation
    Huang, Zhenhua
    Lin, Zuorui
    Gong, Zheng
    Chen, Yunwen
    Tang, Yong
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (09) : 5902 - 5923
  • [23] Channel-level Matching Knowledge Distillation for object detectors via MSE
    Jiang, Zetao
    Huang, Qinyang
    Zhang, Huijuan
    PATTERN RECOGNITION LETTERS, 2024, 179 : 52 - 57
  • [24] CDFKD-MFS: Collaborative Data-Free Knowledge Distillation via Multi-Level Feature Sharing
    Hao, Zhiwei
    Luo, Yong
    Wang, Zhi
    Hu, Han
    An, Jianping
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 4262 - 4274
  • [25] A General Knowledge Distillation Framework for Counterfactual Recommendation via Uniform Data
    Liu, Dugang
    Cheng, Pengxiang
    Dong, Zhenhua
    He, Xiuqiang
    Pan, Weike
    Ming, Zhong
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 831 - 840
  • [26] Enhancing action recognition from low-quality skeleton data via part-level knowledge distillation
    Liu, Cuiwei
    Jiang, Youzhi
    Du, Chong
    Li, Zhaokui
    SIGNAL PROCESSING, 2024, 221
  • [27] Multi-level knowledge distillation via dynamic decision boundaries exploration and exploitation
    Tao, Ze
    Li, Haowei
    Zhang, Jian
    Zhang, Shichao
    INFORMATION FUSION, 2024, 112
  • [28] CBKG-DTI: Multi-Level Knowledge Distillation and Biomedical Knowledge Graph for Drug-Target Interaction Prediction
    Zhao, Xiaosa
    Wang, Qixian
    Zhang, Ye
    He, Chenglong
    Yin, Minghao
    Zhao, Xiaowei
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (03) : 2284 - 2296
  • [29] Incremental event detection via an improved knowledge distillation based model
    Lin, Yi
    Xu, Changhua
    Yu, Hang
    Tian, Pinzhuo
    Luo, Xiangfeng
    NEUROCOMPUTING, 2023, 551
  • [30] Efficient Neural Data Compression for Machine Type Communications via Knowledge Distillation
    Hussien, Mostafa
    Xu, Yi Tian
    Wu, Di
    Liu, Xue
    Dudek, Gregory
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 1169 - 1174