Progressive multi-level distillation learning for pruning network

被引:0
|
作者
Ruiqing Wang
Shengmin Wan
Wu Zhang
Chenlu Zhang
Yu Li
Shaoxiang Xu
Lifu Zhang
Xiu Jin
Zhaohui Jiang
Yuan Rao
机构
[1] Anhui Agricultural University,School of Information and Computer
[2] Anhui Agriculture University,Anhui Province Key Laboratory of Smart Agricultural Technology and Equipment
来源
关键词
Deep neural network; Model compression; Network pruning; Knowledge distillation;
D O I
暂无
中图分类号
学科分类号
摘要
Although the classification method based on the deep neural network has achieved excellent results in classification tasks, it is difficult to apply to real-time scenarios because of high memory footprints and prohibitive inference times. Compared to unstructured pruning, structured pruning techniques can reduce the computation cost of the model runtime more effectively, but inevitably reduces the precision of the model. Traditional methods use fine tuning to restore model damage performance. However, there is still a large gap between the pruned model and the original one. In this paper, we use progressive multi-level distillation learning to compensate for the loss caused by pruning. Pre-pruning and post-pruning networks serve as the teacher and student networks. The proposed approach utilizes the complementary properties of structured pruning and knowledge distillation, which allows the pruned network to learn the intermediate and output representations of the teacher network, thus reducing the influence of the model subject to pruning. Experiments demonstrate that our approach performs better on CIFAR-10, CIFAR-100, and Tiny-ImageNet datasets with different pruning rates. For instance, GoogLeNet can achieve near lossless pruning on the CIFAR-10 dataset with 60% pruning. Moreover, this paper also proves that using the proposed distillation learning method during the pruning process achieves more significant performance gains than after completing the pruning.
引用
收藏
页码:5779 / 5791
页数:12
相关论文
共 50 条
  • [41] MULAN: Multi-Level Adaptive Network Filter
    Tzur-David, Shimrit
    Dolev, Danny
    Anker, Tal
    SECURITY AND PRIVACY IN COMMUNICATION NETWORKS, 2009, 19 : 71 - 90
  • [42] Ensemble relation network with multi-level measure
    Xiaoxu L.
    Jie C.
    Xue Q.
    Journal of China Universities of Posts and Telecommunications, 2022, 29 (03): : 15 - 24
  • [43] Products exchange in a multi-level distribution network
    Benfriha, Abdennour Ilyas
    Triqui-Sari, Lamia
    Bougloula, Aimed Eddine
    Bennekrouf, Mohammed
    2019 INTERNATIONAL COLLOQUIUM ON LOGISTICS AND SUPPLY CHAIN MANAGEMENT (LOGISTIQUA), 2019,
  • [44] Multi-level perception fusion dehazing network
    Wu, Xiaohua
    Li, Zenglu
    Guo, Xiaoyu
    Xiang, Songyang
    Zhang, Yao
    PLOS ONE, 2023, 18 (10):
  • [45] Algorithms for a multi-level network optimization problem
    Cruz, FRB
    Smith, JM
    Mateus, GR
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 1999, 118 (01) : 164 - 180
  • [46] An approach to flexible multi-level network design
    Cheushev, V
    Kolodziejczyk, J
    Luba, T
    Moraga, C
    Sapiecha, P
    Shmerko, V
    Yanushkevich, S
    SYSTEM-ON-CHIP FOR REAL-TIME APPLICATIONS, 2003, : 214 - 223
  • [47] SYLON DREAM - A MULTI-LEVEL NETWORK SYNTHESIZER
    CHEN, KC
    MUROGA, S
    1989 IEEE INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN: DIGEST OF TECHNICAL PAPERS, 1989, : 552 - 555
  • [48] Multi-level disentanglement graph neural network
    Wu, Lirong
    Lin, Haitao
    Xia, Jun
    Tan, Cheng
    Li, Stan Z.
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (11): : 9087 - 9101
  • [49] Multi-level disentanglement graph neural network
    Lirong Wu
    Haitao Lin
    Jun Xia
    Cheng Tan
    Stan Z. Li
    Neural Computing and Applications, 2022, 34 : 9087 - 9101
  • [50] Ensemble relation network with multi-level measure
    Li Xiaoxu
    Qu Xue
    Cao Jie
    The Journal of China Universities of Posts and Telecommunications, 2022, 29 (03) : 15 - 24