Quantization Robust Pruning With Knowledge Distillation

被引:6
作者
Kim, Jangho [1 ]
机构
[1] Kookmin Univ, Coll Comp Sci, Sch Artificial Intelligence, Seoul 02707, South Korea
关键词
Quantization (signal); Computational modeling; Convolutional neural networks; Knowledge engineering; Robustness; Performance evaluation; Neural networks; network quantization; knowledge distillation; network pruning;
D O I
10.1109/ACCESS.2023.3257864
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To resolve the problem that deep neural networks (DNN) require a large number of network parameters, many researchers have sought to compress the network. Network pruning, quantization and knowledge distillation have been studied for this purpose. Considering realistic scenarios such as deploying DNN on the resource constraint device where the network uploaded in the device performs wells in various bit-widths without re-training and the network with reasonable performance, we propose quantization robust pruning with knowledge distillation (QRPK) method. In QRPK, model weights are divided into essential weigths and inessential weights based on their magnitude value. Then, QRPK trains the quantization robustness model with a high pruning ratio by making the distribution of essential weights as a quantization friendly distribution. We conducted experiments on CIFAR-10 and CIFAR-100 to verify the effectiveness of QRPK and a QRPK trained model performs well in various bit-width, as it designed by pruning, quantization robustness and knowledge distillation.
引用
收藏
页码:26419 / 26426
页数:8
相关论文
共 27 条
  • [1] Alizadeh M., 2020, 8 INT C LEARNING REP, P1
  • [2] Banner R, 2019, ADV NEUR IN, V32
  • [3] Chen Tianyi, 2021, ADV NEURAL INFORM PR, V34
  • [4] Chung I, 2020, PR MACH LEARN RES, V119
  • [5] Frankle J., 2018, arXiv
  • [6] Differentiable Soft Quantization: Bridging Full-Precision and Low-Bit Neural Networks
    Gong, Ruihao
    Liu, Xianglong
    Jiang, Shenghu
    Li, Tianxiang
    Hu, Peng
    Lin, Jiazhen
    Yu, Fengwei
    Yan, Junjie
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 4851 - 4860
  • [7] Han S, 2015, ADV NEUR IN, V28
  • [8] Hinton G, 2015, Arxiv, DOI [arXiv:1503.02531, DOI 10.48550/ARXIV.1503.02531]
  • [9] Hu P, 2021, AAAI CONF ARTIF INTE, V35, P7780
  • [10] Esser SK, 2020, Arxiv, DOI arXiv:1902.08153