Progressive multi-level distillation learning for pruning network

被引:0
|
作者
Ruiqing Wang
Shengmin Wan
Wu Zhang
Chenlu Zhang
Yu Li
Shaoxiang Xu
Lifu Zhang
Xiu Jin
Zhaohui Jiang
Yuan Rao
机构
[1] Anhui Agricultural University,School of Information and Computer
[2] Anhui Agriculture University,Anhui Province Key Laboratory of Smart Agricultural Technology and Equipment
来源
关键词
Deep neural network; Model compression; Network pruning; Knowledge distillation;
D O I
暂无
中图分类号
学科分类号
摘要
Although the classification method based on the deep neural network has achieved excellent results in classification tasks, it is difficult to apply to real-time scenarios because of high memory footprints and prohibitive inference times. Compared to unstructured pruning, structured pruning techniques can reduce the computation cost of the model runtime more effectively, but inevitably reduces the precision of the model. Traditional methods use fine tuning to restore model damage performance. However, there is still a large gap between the pruned model and the original one. In this paper, we use progressive multi-level distillation learning to compensate for the loss caused by pruning. Pre-pruning and post-pruning networks serve as the teacher and student networks. The proposed approach utilizes the complementary properties of structured pruning and knowledge distillation, which allows the pruned network to learn the intermediate and output representations of the teacher network, thus reducing the influence of the model subject to pruning. Experiments demonstrate that our approach performs better on CIFAR-10, CIFAR-100, and Tiny-ImageNet datasets with different pruning rates. For instance, GoogLeNet can achieve near lossless pruning on the CIFAR-10 dataset with 60% pruning. Moreover, this paper also proves that using the proposed distillation learning method during the pruning process achieves more significant performance gains than after completing the pruning.
引用
收藏
页码:5779 / 5791
页数:12
相关论文
共 50 条
  • [21] A multi-level probabilistic neural network
    Zong, Ning
    Hong, Xia
    ADVANCES IN NEURAL NETWORKS - ISNN 2007, PT 2, PROCEEDINGS, 2007, 4492 : 516 - +
  • [22] Multi-Level network security architecture
    Luo, J.Z.
    Liu, Y.Y.
    2001, Chinese Academy of Sciences (12):
  • [23] MLAN: Multi-Level Attention Network
    Qin, Peinuan
    Wang, Qinxuan
    Zhang, Yue
    Wei, Xueyao
    Gao, Meiguo
    IEEE ACCESS, 2022, 10 : 105437 - 105446
  • [24] Progressive engineering changes in multi-level product structures
    Ho, CJ
    Li, J
    DECISION SCIENCES INSTITUTE, 1997 ANNUAL MEETING, PROCEEDINGS, VOLS 1-3, 1997, : 1408 - 1410
  • [25] Progressive engineering changes in multi-level product structures
    Ho, CJ
    Li, J
    OMEGA-INTERNATIONAL JOURNAL OF MANAGEMENT SCIENCE, 1997, 25 (05): : 585 - 594
  • [26] Hierarchical Bayesian learning framework for multi-level modeling using multi-level data
    Jia, Xinyu
    Papadimitriou, Costas
    MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2022, 179
  • [27] Multi-level graph contrastive learning
    Shao, Pengpeng
    Tao, Jianhua
    NEUROCOMPUTING, 2024, 570
  • [28] Incremental learning with multi-level adaptation
    Bouchachia, Abdelhamid
    NEUROCOMPUTING, 2011, 74 (11) : 1785 - 1799
  • [29] NESTED LEARNING FOR MULTI-LEVEL CLASSIFICATION
    Achddou, Raphael
    Di Martino, J. Matias
    Sapiro, Guillermo
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2815 - 2819
  • [30] Multiresolution Knowledge Distillation and Multi-level Fusion for Defect Detection
    Xie, Huosheng
    Xiao, Yan
    GREEN, PERVASIVE, AND CLOUD COMPUTING, GPC 2022, 2023, 13744 : 178 - 191