Model Selection - Knowledge Distillation Framework for Model Compression

被引:0
|
作者
Chen, Renhai [1 ]
Yuan, Shimin [1 ]
Wang, Shaobo [1 ]
Li, Zhenghan [1 ]
Xing, Meng [1 ]
Feng, Zhiyong [1 ]
机构
[1] Tianjin Univ, Shenzhen Res Inst, Coll Intelligence & Comp, Tianjin, Peoples R China
来源
2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021) | 2021年
基金
中国国家自然科学基金;
关键词
model selection; model compression; knowledge distillation;
D O I
10.1109/SSCI50451.2021.9659861
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The significant increase in the computation and parameter storage costs of CNNs promotes its development in various applications and restricts its deployment in edge devices as well. Therefore, many neural network pruning methods has been proposed for neural network compression and acceleration. However, there are two major limitations to these methods: First, prevailing methods usually design single pruning criteria for the primitive network and fail to consider the diversity of potential optimal sub-network structure. Second, these methods utilize traditional training method to train the sub-network, which is not enough to develop the expression ability of the sub-network under the current task.In this paper, we propose Model Selection - Knowledge Distillation (MS-KD) framework to solve the above problems. Specifically, we develop multiple pruning criteria for the primitive network, and the potential optimal structure is obtained through model selection.Furthermore, instead of traditional training methods, we use knowledge distillation to train the learned sub-network and make full use of the structure advantages of the sub-network.To validate our approach, we conduct extensive experiments on prevalent image classification datasets.The results demonstrate that our MS-KD framework outperforms the existing methods under a wide range of data sets, models, and inference costs.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] How to Train the Teacher Model for Effective Knowledge Distillation
    Hamidi, Shayan Mohajer
    Deng, Xizhen
    Tan, Renhao
    Ye, Linfeng
    Salamah, Ahmed Hussein
    COMPUTER VISION - ECCV 2024, PT LXXXIX, 2025, 15147 : 1 - 18
  • [42] A Video Question Answering Model Based on Knowledge Distillation
    Shao, Zhuang
    Wan, Jiahui
    Zong, Linlin
    INFORMATION, 2023, 14 (06)
  • [43] Dual model knowledge distillation for industrial anomaly detection
    Thomine, Simon
    Snoussi, Hichem
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (03)
  • [44] Knowledge Distillation for In-Memory Keyword Spotting Model
    Song, Zeyang
    Liu, Qi
    Yang, Qu
    Li, Haizhou
    INTERSPEECH 2022, 2022, : 4128 - 4132
  • [45] KDRSFL: A knowledge distillation resistance transfer framework for defending model inversion attacks in split federated learning
    Chen, Renlong
    Xia, Hui
    Wang, Kai
    Xu, Shuo
    Zhang, Rui
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2025, 166
  • [46] A Lightweight Malware Detection Model Based on Knowledge Distillation
    Miao, Chunyu
    Kou, Liang
    Zhang, Jilin
    Dong, Guozhong
    MATHEMATICS, 2024, 12 (24)
  • [47] Intelligent reconstruction algorithm of hydrogen-fueled scramjet combustor flow based on knowledge distillation model compression
    Tian, Ye
    Wang, Gang
    Deng, Xue
    Guo, Mingming
    Ren, Hu
    Li, Linjing
    Chen, Erda
    Zhang, Hua
    Le, Jialing
    INTERNATIONAL JOURNAL OF HYDROGEN ENERGY, 2024, 49 : 1278 - 1291
  • [48] Parallel Blockwise Knowledge Distillation for Deep Neural Network Compression
    Blakeney, Cody
    Li, Xiaomin
    Yan, Yan
    Zong, Ziliang
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2021, 32 (07) : 1765 - 1776
  • [49] Incremental Deep Learning Method for Object Detection Model Based on Knowledge Distillation
    Fang W.
    Chen A.
    Meng N.
    Cheng H.
    Wang Q.
    Gongcheng Kexue Yu Jishu/Advanced Engineering Sciences, 2022, 54 (06): : 59 - 66
  • [50] Model compression through distillation with cross-layer integrated guidance at word level
    Li, Guiyu
    Zheng, Shang
    Zou, Haitao
    Yu, Hualong
    Gao, Shang
    NEUROCOMPUTING, 2025, 619