Progressive multi-level distillation learning for pruning network

被引:0
|
作者
Ruiqing Wang
Shengmin Wan
Wu Zhang
Chenlu Zhang
Yu Li
Shaoxiang Xu
Lifu Zhang
Xiu Jin
Zhaohui Jiang
Yuan Rao
机构
[1] Anhui Agricultural University,School of Information and Computer
[2] Anhui Agriculture University,Anhui Province Key Laboratory of Smart Agricultural Technology and Equipment
来源
关键词
Deep neural network; Model compression; Network pruning; Knowledge distillation;
D O I
暂无
中图分类号
学科分类号
摘要
Although the classification method based on the deep neural network has achieved excellent results in classification tasks, it is difficult to apply to real-time scenarios because of high memory footprints and prohibitive inference times. Compared to unstructured pruning, structured pruning techniques can reduce the computation cost of the model runtime more effectively, but inevitably reduces the precision of the model. Traditional methods use fine tuning to restore model damage performance. However, there is still a large gap between the pruned model and the original one. In this paper, we use progressive multi-level distillation learning to compensate for the loss caused by pruning. Pre-pruning and post-pruning networks serve as the teacher and student networks. The proposed approach utilizes the complementary properties of structured pruning and knowledge distillation, which allows the pruned network to learn the intermediate and output representations of the teacher network, thus reducing the influence of the model subject to pruning. Experiments demonstrate that our approach performs better on CIFAR-10, CIFAR-100, and Tiny-ImageNet datasets with different pruning rates. For instance, GoogLeNet can achieve near lossless pruning on the CIFAR-10 dataset with 60% pruning. Moreover, this paper also proves that using the proposed distillation learning method during the pruning process achieves more significant performance gains than after completing the pruning.
引用
收藏
页码:5779 / 5791
页数:12
相关论文
共 50 条
  • [1] Progressive multi-level distillation learning for pruning network
    Wang, Ruiqing
    Wan, Shengmin
    Zhang, Wu
    Zhang, Chenlu
    Li, Yu
    Xu, Shaoxiang
    Zhang, Lifu
    Jin, Xiu
    Jiang, Zhaohui
    Rao, Yuan
    COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (05) : 5779 - 5791
  • [2] Multi-level Logit Distillation
    Jin, Ying
    Wang, Jiaqi
    Lin, Dahua
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24276 - 24285
  • [3] Multi-Level Progressive Reinforcement Learning for Control Policy in Physical Simulations
    Wu, Kefei
    He, Xuming
    Wang, Yang
    Liu, Xiaopei
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 9502 - 9508
  • [4] Multi-level progressive transfer learning for cervical cancer dose prediction
    Wen, Lu
    Xiao, Jianghong
    Zeng, Jie
    Zu, Chen
    Wu, Xi
    Zhou, Jiliu
    Peng, Xingchen
    Wang, Yan
    PATTERN RECOGNITION, 2023, 141
  • [5] Multi-Level Progressive Learning for Unsupervised Vehicle Re-Identification
    He, Zhijun
    Zhao, Hongbo
    Wang, Jianrong
    Feng, Wenquan
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (04) : 4357 - 4371
  • [6] Progressive Raising in Multi-level IR
    Chelini, Lorenzo
    Drebes, Andi
    Zinenko, Oleksandr
    Cohen, Albert
    Vasilache, Nicolas
    Grosser, Tobias
    Corporaal, Henk
    CGO '21: PROCEEDINGS OF THE 2021 IEEE/ACM INTERNATIONAL SYMPOSIUM ON CODE GENERATION AND OPTIMIZATION (CGO), 2021, : 15 - 26
  • [7] Multi-level network Lasso for multi-task personalized learning
    Wang, Jiankun
    Fei, Luhuan
    Sun, Lu
    PATTERN RECOGNITION, 2025, 161
  • [8] Multi-level graph learning network for hyperspectral image classification
    Wan, Sheng
    Pan, Shirui
    Zhong, Shengwei
    Yang, Jie
    Yang, Jian
    Zhan, Yibing
    Gong, Chen
    PATTERN RECOGNITION, 2022, 129
  • [9] Adaptive multi-teacher multi-level knowledge distillation
    Liu, Yuang
    Zhang, Wei
    Wang, Jun
    NEUROCOMPUTING, 2020, 415 : 106 - 113
  • [10] Multi-level features progressive refinement and edge enhancement network for image dehazing
    Fu Y.
    Yin S.
    Deng Z.
    Wang Y.
    Hu S.
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2022, 30 (09): : 1091 - 1100