Spirit Distillation: A Model Compression Method with Multi-domain Knowledge Transfer

被引:6
|
作者
Wu, Zhiyuan [1 ]
Jiang, Yu [1 ,2 ]
Zhao, Minghao [1 ]
Cui, Chupeng [1 ]
Yang, Zongmin [1 ]
Xue, Xinhui [1 ]
Qi, Hong [1 ,2 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Changchun, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Computat & Knowledge Engn, Changchun, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT I | 2021年 / 12815卷
基金
中国国家自然科学基金;
关键词
Knowledge transfer; Knowledge distillation; Multi-domain; Model compression; Few-shot learning;
D O I
10.1007/978-3-030-82136-4_45
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent applications pose requirements of both cross-domain knowledge transfer and model compression to machine learning models due to insufficient training data and limited computational resources. In this paper, we propose a new knowledge distillation model, named Spirit Distillation (SD), which is a model compression method with multi-domain knowledge transfer. The compact student network mimics out a representation equivalent to the front part of the teacher network, through which the general knowledge can be transferred from the source domain (teacher) to the target domain (student). To further improve the robustness of the student, we extend SD to Enhanced Spirit Distillation (ESD) in exploiting a more comprehensive knowledge by introducing the proximity domainwhich is similar to the target domain for feature extraction. Persuasive experiments are conducted on Cityscapes semantic segmentation with the prior knowledge transferred fromCOCO2017 and KITTI. Results demonstrate that our method can boost mIOU and high-precision accuracy by 1.4% and 8.2% respectively with 78.2% segmentation variance, and can gain a precise compact network with only 41.8% FLOPs.
引用
收藏
页码:553 / 565
页数:13
相关论文
共 50 条
  • [31] NFFKD: A Knowledge Distillation Method Based on Normalized Feature Fusion Model
    Wang, Zihan
    Xie, Junwei
    Yao, Zhiping
    Kuang, Xu
    Gao, Qinquan
    Tong, Tong
    2022 IEEE THE 5TH INTERNATIONAL CONFERENCE ON BIG DATA AND ARTIFICIAL INTELLIGENCE (BDAI 2022), 2022, : 111 - 116
  • [32] A Dynamic Multi-domain Access Control Model in Cloud Computing
    Xiong, Dapeng
    Zou, Peng
    Cai, Jun
    He, Jun
    SECURITY IN COMPUTING AND COMMUNICATIONS (SSCC 2015), 2015, 536 : 3 - 12
  • [33] Data-Free Ensemble Knowledge Distillation for Privacy-conscious Multimedia Model Compression
    Hao, Zhiwei
    Luo, Yong
    Hu, Han
    An, Jianping
    Wen, Yonggang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1803 - 1811
  • [34] KDTM: Multi-Stage Knowledge Distillation Transfer Model for Long-Tailed DGA Detection
    Fan, Baoyu
    Ma, Han
    Liu, Yue
    Yuan, Xiaochen
    Ke, Wei
    MATHEMATICS, 2024, 12 (05)
  • [35] A method of assessing the time-variant value of multi-domain architectures
    Ford, Thomas
    Meyer, David
    Colombi, John
    Scheller, Brian
    Palmer, Cody
    JOURNAL OF DEFENSE MODELING AND SIMULATION-APPLICATIONS METHODOLOGY TECHNOLOGY-JDMS, 2018, 15 (04): : 471 - 481
  • [36] A new multi-domain/multi-resolution method for large-eddy simulation
    Quéméré, P
    Sagaut, P
    Couailler, V
    INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN FLUIDS, 2001, 36 (04) : 391 - 416
  • [37] A Hybrid EEG-Based Stress State Classification Model Using Multi-Domain Transfer Entropy and PCANet
    Dong, Yuefang
    Xu, Lin
    Zheng, Jian
    Wu, Dandan
    Li, Huanli
    Shao, Yongcong
    Shi, Guohua
    Fu, Weiwei
    BRAIN SCIENCES, 2024, 14 (06)
  • [38] A multi-domain spectral method for time-fractional differential equations
    Chen, Feng
    Xu, Qinwu
    Hesthaven, Jan S.
    JOURNAL OF COMPUTATIONAL PHYSICS, 2015, 293 : 157 - 172
  • [39] Incremental Deep Learning Method for Object Detection Model Based on Knowledge Distillation
    Fang W.
    Chen A.
    Meng N.
    Cheng H.
    Wang Q.
    Gongcheng Kexue Yu Jishu/Advanced Engineering Sciences, 2022, 54 (06): : 59 - 66
  • [40] Single-Model Multi-domain Dialogue Management with Deep Learning
    Papangelis, Alexandros
    Stylianou, Yannis
    ADVANCED SOCIAL INTERACTION WITH AGENTS, 2019, 510 : 71 - 77