Knowledge Distillation via Token-Level Relationship Graph Based on the Big Data Technologies

被引:3
|
作者
Zhang, Shuoxi [1 ]
Liu, Hanpeng [1 ]
He, Kun [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, 1037 Luoyu Rd, Wuhan 430074, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge distillation; Graph representation; Graph-based distillation; Big data technology; NEURAL-NETWORKS;
D O I
10.1016/j.bdr.2024.100438
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the big data era, characterized by vast volumes of complex data, the efficiency of machine learning models is of utmost importance, particularly in the context of intelligent agriculture. Knowledge distillation (KD), a technique aimed at both model compression and performance enhancement, serves as a pivotal solution by distilling the knowledge from an elaborate model (teacher) to a lightweight, compact counterpart (student). However, the true potential of KD has not been fully explored. Existing approaches primarily focus on transferring instancelevel information by big data technologies, overlooking the valuable information embedded in token-level relationships, which may be particularly affected by the long-tail effects. To address the above limitations, we propose a novel method called Knowledge Distillation with Token-level Relationship Graph (TRG) that leverages token-wise relationships to enhance the performance of knowledge distillation. By employing TRG, the student model can effectively emulate higher-level semantic information from the teacher model, resulting in improved performance and mobile-friendly efficiency. To further enhance the learning process, we introduce a dynamic temperature adjustment strategy, which encourages the student model to capture the topology structure of the teacher model more effectively. We conduct experiments to evaluate the effectiveness of the proposed method against several state-of-the-art approaches. Empirical results demonstrate the superiority of TRG across various visual tasks, including those involving imbalanced data. Our method consistently outperforms the existing baselines, establishing a new state-of-the-art performance in the field of KD based on big data technologies.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Token-Level Ensemble Distillation for Grapheme-to-Phoneme Conversion
    Sun, Hao
    Tan, Xu
    Gan, Jun-Wei
    Liu, Hongzhi
    Zhao, Sheng
    Qin, Tao
    Liu, Tie-Yan
    INTERSPEECH 2019, 2019, : 2115 - 2119
  • [2] The Devil is in the Data: Learning Fair Graph Neural Networks via Partial Knowledge Distillation
    Zhu, Yuchang
    Li, Jintang
    Chen, Liang
    Zheng, Zibin
    PROCEEDINGS OF THE 17TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2024, 2024, : 1012 - 1021
  • [3] Few-Shot Graph Anomaly Detection via Dual-Level Knowledge Distillation
    Li, Xuan
    Cheng, Dejie
    Zhang, Luheng
    Zhang, Chengfang
    Feng, Ziliang
    ENTROPY, 2025, 27 (01)
  • [4] Generalized Knowledge Distillation via Relationship Matching
    Ye, Han-Jia
    Lu, Su
    Zhan, De-Chuan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 1817 - 1834
  • [5] Movie Big Data Intelligent Recommendation System Based on Knowledge Graph
    Qiu, Gang
    Guo, Yanli
    19TH IEEE INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED PROCESSING WITH APPLICATIONS (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2021), 2021, : 539 - 546
  • [6] A transformer based visual tracker with restricted token interaction and knowledge distillation
    Liu, Nian
    Zhang, Yi
    KNOWLEDGE-BASED SYSTEMS, 2025, 307
  • [7] Deep Graph-level Anomaly Detection by Glocal Knowledge Distillation
    Ma, Rongrong
    Pang, Guansong
    Chen, Ling
    van den Hengel, Anton
    WSDM'22: PROCEEDINGS OF THE FIFTEENTH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2022, : 704 - 714
  • [8] Trace Data Analytics with Knowledge Distillation DM: Big Data Management and Mining
    Lee, Janghwan
    Xiong, Wei
    Jang, Wonhyouk
    2020 31ST ANNUAL SEMI ADVANCED SEMICONDUCTOR MANUFACTURING CONFERENCE (ASMC), 2020,
  • [9] Difficulty level-based knowledge distillation
    Ham, Gyeongdo
    Cho, Yucheol
    Lee, Jae-Hyeok
    Kang, Minchan
    Choi, Gyuwon
    Kim, Daeshik
    NEUROCOMPUTING, 2024, 606
  • [10] Grand: A Fast and Accurate Graph Retrieval Framework via Knowledge Distillation
    Lan, Lin
    Wang, Pinghui
    Shi, Rui
    Liu, Tingqing
    Zeng, Juxiang
    Sun, Feiyang
    Ren, Yang
    Tao, Jing
    Guan, Xiaohong
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 1639 - 1648