Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

被引:6
|
作者
Wu, Zhiyuan [1 ]
Jiang, Yu [1 ,2 ]
Zhao, Minghao [1 ]
Cui, Chupeng [1 ]
Yang, Zongmin [1 ]
Xue, Xinhui [1 ]
Qi, Hong [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I | 2021年 / 12815卷
基金
中国国家自然科学基金;
关键词
Knowledge transfer; Knowledge distillation; Multi-domain; Model compression; Few-shot learning;
D O I
10.1007/978-3-030-82136-4_45
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domainwhich is similar to the target domain for feature extraction. Persuasive experiments are conducted on Cityscapes semantic segmentation with the prior knowledge transferred fromCOCO2017 and KITTI. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.
引用
收藏
页码:553 / 565
页数:13
相关论文
共 50 条
  • [21] DISCOVER THE EFFECTIVE STRATEGY FOR FACE RECOGNITION MODEL COMPRESSION BY IMPROVED KNOWLEDGE DISTILLATION
    Wang, Mengjiao
    Liu, Rujie
    Abe, Narishige
    Uchida, Hidetsugu
    Matsunami, Tomoaki
    Yamada, Shigefumi
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2416 - 2420
  • [22] Matching the Ideal Pruning Method with Knowledge Distillation for Optimal Compression
    Malihi, Leila
    Heidemann, Gunther
    APPLIED SYSTEM INNOVATION, 2024, 7 (04)
  • [23] Research on the Construction of an Efficient and Lightweight Online Detection Method for Tiny Surface Defects through Model Compression and Knowledge Distillation
    Chen, Qipeng
    Xiong, Qiaoqiao
    Huang, Haisong
    Tang, Saihong
    Liu, Zhenghong
    ELECTRONICS, 2024, 13 (02)
  • [24] Multi-Domain Lifelong Visual Question Answering via Self-Critical Distillation
    Lao, Mingrui
    Pu, Nan
    Liu, Yu
    Zhong, Zhun
    Bakker, Erwin M.
    Sebe, Nicu
    Lew, Michael S.
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4747 - 4758
  • [25] PQK: Model Compression via Pruning, Quantization, and Knowledge Distillation
    Kim, Jangho
    Chang, Simyung
    Kwak, Nojun
    INTERSPEECH 2021, 2021, : 4568 - 4572
  • [26] Cross-lingual distillation for domain knowledge transfer with sentence transformers
    Piperno, Ruben
    Bacco, Luca
    Dell'Orletta, Felice
    Merone, Mario
    Pecchia, Leandro
    KNOWLEDGE-BASED SYSTEMS, 2025, 311
  • [27] SmipRef: An efficient method for multi-domain clock skew scheduling
    Zhi, Yanling
    Luk, Wai-Shing
    Zhou, Hai
    Zeng, Xuan
    INTEGRATION-THE VLSI JOURNAL, 2013, 46 (04) : 392 - 403
  • [28] Multi-Domain Transfer Learning for Early Diagnosis of Alzheimer’s Disease
    Bo Cheng
    Mingxia Liu
    Dinggang Shen
    Zuoyong Li
    Daoqiang Zhang
    Neuroinformatics, 2017, 15 : 115 - 132
  • [29] Multi-Domain Transfer Learning for Early Diagnosis of Alzheimer's Disease
    Cheng, Bo
    Liu, Mingxia
    Shen, Dinggang
    Li, Zuoyong
    Zhang, Daoqiang
    NEUROINFORMATICS, 2017, 15 (02) : 115 - 132
  • [30] A novel multi-domain machine reading comprehension model with domain interference mitigation
    Zhou, Chulun
    Wang, Zhihao
    He, Shaojie
    Zhang, Haiying
    Su, Jinsong
    NEUROCOMPUTING, 2022, 500 : 791 - 798