Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

被引:6
作者
Wu, Zhiyuan [1 ]
Jiang, Yu [1 ,2 ]
Zhao, Minghao [1 ]
Cui, Chupeng [1 ]
Yang, Zongmin [1 ]
Xue, Xinhui [1 ]
Qi, Hong [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I | 2021年 / 12815卷
基金
中国国家自然科学基金;
关键词
Knowledge transfer; Knowledge distillation; Multi-domain; Model compression; Few-shot learning;
D O I
10.1007/978-3-030-82136-4_45
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domainwhich is similar to the target domain for feature extraction. Persuasive experiments are conducted on Cityscapes semantic segmentation with the prior knowledge transferred fromCOCO2017 and KITTI. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.
引用
收藏
页码:553 / 565
页数:13
相关论文
共 50 条
  • [41] A secure and efficient multi-domain data sharing model on consortium chain
    Wenbo Zhang
    Xiaotong Huo
    Zhenshan Bao
    The Journal of Supercomputing, 2023, 79 : 8538 - 8582
  • [42] Trust-based Access Control Model in Multi-domain Environment
    Zhang Qikun
    Wang Ruifang
    Qu Jiaqing
    Gan Yong
    Zheng Jun
    INTERNATIONAL JOURNAL OF SECURITY AND ITS APPLICATIONS, 2014, 8 (05): : 149 - 160
  • [43] A secure and efficient multi-domain data sharing model on consortium chain
    Zhang, Wenbo
    Huo, Xiaotong
    Bao, Zhenshan
    JOURNAL OF SUPERCOMPUTING, 2023, 79 (08) : 8538 - 8582
  • [44] Ensemble Compressed Language Model Based on Knowledge Distillation and Multi-Task Learning
    Xiang, Kun
    Fujii, Akihiro
    2022 7TH INTERNATIONAL CONFERENCE ON BUSINESS AND INDUSTRIAL RESEARCH (ICBIR2022), 2022, : 72 - 77
  • [45] End-to-end model compression via pruning and knowledge distillation for lightweight image super resolution
    Yanzhe Wang
    Yizhen Wang
    Avinash Rohra
    Baoqun Yin
    Pattern Analysis and Applications, 2025, 28 (2)
  • [46] Model compression via pruning and knowledge distillation for person re-identification
    Xie, Haonan
    Jiang, Wei
    Luo, Hao
    Yu, Hongyan
    JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2021, 12 (02) : 2149 - 2161
  • [47] Ontology-Driven Knowledge Modeling and Reasoning for Multi-domain System Architecting and Configuration
    Petnga, Leonard
    RECENT TRENDS AND ADVANCES IN MODEL BASED SYSTEMS ENGINEERING, 2022, : 229 - 239
  • [48] KNOWLEDGE TRANSFER AND MODEL COMPRESSION FOR MISALIGNED BUILDING LABELS
    Neupane, Bipul
    Aryal, Jagannath
    Rajabifard, Abbas
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 3632 - 3635
  • [49] Prediction method of furnace temperature based on transfer learning and knowledge distillation
    Zhai N.
    Zhou X.F.
    Li S.
    Shi H.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2022, 28 (06): : 1860 - 1869
  • [50] Joint structured pruning and dense knowledge distillation for efficient transformer model compression
    Cui, Baiyun
    Li, Yingming
    Zhang, Zhongfei
    NEUROCOMPUTING, 2021, 458 : 56 - 69