Model Compression Algorithm via Reinforcement Learning and Knowledge Distillation

被引:1
作者
Liu, Botao [1 ]
Hu, Bing-Bing [1 ]
Zhao, Ming [1 ]
Peng, Sheng-Lung [2 ]
Chang, Jou-Ming [3 ]
Tsoulos, Ioannis G.
机构
[1] Yangtze Univ, Sch Comp Sci, Jingzhou 434025, Peoples R China
[2] Natl Taipei Univ Business, Dept Creat Technol & Prod Design, Taoyuan 10051, Taiwan
[3] Natl Taipei Univ Business, Inst Informat & Decis Sci, Taipei 10051, Taiwan
关键词
model compression; reinforcement learning; knowledge distillation; attention mechanism; automatic pruning; network efficiency;
D O I
10.3390/math11224589
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Traditional model compression techniques are dependent on handcrafted features and require domain experts, with a tradeoff between model size, speed, and accuracy. This study proposes a new approach toward resolving model compression problems. Our approach combines reinforcement-learning-based automated pruning and knowledge distillation to improve the pruning of unimportant network layers and the efficiency of the compression process. We introduce a new state quantity that controls the size of the reward and an attention mechanism that reinforces useful features and attenuates useless features to enhance the effects of other features. The experimental results show that the proposed model is superior to other advanced pruning methods in terms of the computation time and accuracy on CIFAR-100 and ImageNet dataset, where the accuracy is approximately 3% higher than that of similar methods with shorter computation times.
引用
收藏
页数:12
相关论文
共 39 条
  • [11] He Y, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2234
  • [12] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
    He, Yihui
    Lin, Ji
    Liu, Zhijian
    Wang, Hanrui
    Li, Li-Jia
    Han, Song
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 815 - 832
  • [13] Hinton G, 2015, Arxiv, DOI [arXiv:1503.02531, DOI 10.48550/ARXIV.1503.02531]
  • [14] Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
  • [15] Densely Connected Convolutional Networks
    Huang, Gao
    Liu, Zhuang
    van der Maaten, Laurens
    Weinberger, Kilian Q.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2261 - 2269
  • [16] Hwang K, 2014, IEEE WRK SIG PRO SYS, P174
  • [17] Kim I., 2020, P IEEE CVF C COMP VI, P9533
  • [18] Kim YD, 2016, Arxiv, DOI arXiv:1511.06530
  • [19] Compositional Convolutional Neural Networks: A Deep Architecture with Innate Robustness to Partial Occlusion
    Kortylewski, Adam
    He, Ju
    Liu, Qing
    Yuille, Alan
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 8937 - 8946
  • [20] ImageNet Classification with Deep Convolutional Neural Networks
    Krizhevsky, Alex
    Sutskever, Ilya
    Hinton, Geoffrey E.
    [J]. COMMUNICATIONS OF THE ACM, 2017, 60 (06) : 84 - 90