Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

被引:6
|
作者
Wu, Zhiyuan [1 ]
Jiang, Yu [1 ,2 ]
Zhao, Minghao [1 ]
Cui, Chupeng [1 ]
Yang, Zongmin [1 ]
Xue, Xinhui [1 ]
Qi, Hong [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I | 2021年 / 12815卷
基金
中国国家自然科学基金;
关键词
Knowledge transfer; Knowledge distillation; Multi-domain; Model compression; Few-shot learning;
D O I
10.1007/978-3-030-82136-4_45
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domainwhich is similar to the target domain for feature extraction. Persuasive experiments are conducted on Cityscapes semantic segmentation with the prior knowledge transferred fromCOCO2017 and KITTI. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.
引用
收藏
页码:553 / 565
页数:13
相关论文
共 50 条
  • [1] A MULTI-DOMAIN KNOWLEDGE TRANSFER METHOD FOR CONCEPTUAL DESIGN COMBINE WITH FBS AND KNOWLEDGE GRAPHA MULTI-DOMAIN KNOWLEDGE TRANSFER METHOD FOR CONCEPTUAL DESIGN COMBINE WITH FBS AND KNOWLEDGE GRAPH
    Lai, Bing
    Zhao, Wu
    Yu, Zeyuan
    Guo, Xin
    Zhang, Kai
    PROCEEDINGS OF ASME 2022 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2022, VOL 2, 2022,
  • [2] Multi-domain Knowledge Distillation via Uncertainty-Matching for End-to-End ASR Models
    Kim, Ho-Gyeong
    Lee, Min-Joong
    Lee, Hoshik
    Kang, Tae Gyoon
    Lee, Jihyun
    Yang, Eunho
    Hwang, Sung Ju
    INTERSPEECH 2021, 2021, : 2531 - 2535
  • [3] Adapter-Based Selective Knowledge Distillation for Federated Multi-Domain Meeting Summarization
    Feng, Xiachong
    Feng, Xiaocheng
    Du, Xiyuan
    Kan, Min-Yen
    Qin, Bing
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3694 - 3708
  • [4] Model Selection - Knowledge Distillation Framework for Model Compression
    Chen, Renhai
    Yuan, Shimin
    Wang, Shaobo
    Li, Zhenghan
    Xing, Meng
    Feng, Zhiyong
    2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
  • [5] Triplet Knowledge Distillation Networks for Model Compression
    Tang, Jialiang
    Jiang, Ning
    Yu, Wenxin
    Wu, Wenqin
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [6] Analysis of Model Compression Using Knowledge Distillation
    Hong, Yu-Wei
    Leu, Jenq-Shiou
    Faisal, Muhamad
    Prakosa, Setya Widyawan
    IEEE ACCESS, 2022, 10 : 85095 - 85105
  • [7] Compression of Acoustic Model via Knowledge Distillation and Pruning
    Li, Chenxing
    Zhu, Lei
    Xu, Shuang
    Gao, Peng
    Xu, Bo
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 2785 - 2790
  • [8] Hierarchical Multi-Attention Transfer for Knowledge Distillation
    Gou, Jianping
    Sun, Liyuan
    Yu, Baosheng
    Wan, Shaohua
    Tao, Dacheng
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2024, 20 (02)
  • [9] Model Compression Algorithm via Reinforcement Learning and Knowledge Distillation
    Liu, Botao
    Hu, Bing-Bing
    Zhao, Ming
    Peng, Sheng-Lung
    Chang, Jou-Ming
    Tsoulos, Ioannis G.
    MATHEMATICS, 2023, 11 (22)
  • [10] Model Compression Based on Knowledge Distillation and Its Application in HRRP
    Chen, Xiaojiao
    An, Zhenyu
    Huang, Liansheng
    He, Shiying
    Wang, Zhen
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 1268 - 1272