REINFORCEMENT LEARNING-BASED LAYER-WISE QUANTIZATION FOR LIGHTWEIGHT DEEP NEURAL NETWORKS

被引:0
|
作者
Jung, Juri [1 ]
Kim, Jonghee [1 ]
Kim, Youngeun [1 ]
Kim, Changick [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Daejeon, South Korea
来源
2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) | 2020年
关键词
Deep neural network; Reinforcement learning; Model compression; Quantization; Embedded system;
D O I
暂无
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
Network quantization has been widely studied to compress the deep neural network in mobile devices. Conventional methods quantize the network parameters of all layers with the same fixed precision, regardless of the number of parameters in each layer. However, quantizing the weights of the layer with many parameters is more effective in reducing the model size. Accordingly, in this paper, we propose a novel mixed-precision quantization method based on reinforcement learning. Specifically, we utilize the number of parameters at each layer as a prior for our framework. By using the accuracy and the bit-width as a reward, the proposed framework determines the optimal quantization policy for each layer. By applying this policy sequentially, we achieve weighted-average 2.97 bits for the VGG-16 model on the CIFAR-10 dataset with no degradation of the accuracy, compared with its full-precision baseline. We also show that our framework can provide an optimal quantization policy for the VGG-Net and the ResNet to minimize the storage while preserving the accuracy.
引用
收藏
页码:3070 / 3074
页数:5
相关论文
共 50 条
  • [1] Collaborative Layer-Wise Discriminative Learning in Deep Neural Networks
    Jin, Xiaojie
    Chen, Yunpeng
    Dong, Jian
    Feng, Jiashi
    Yan, Shuicheng
    COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 : 733 - 749
  • [2] A Layer-Wise Theoretical Framework for Deep Learning of Convolutional Neural Networks
    Huu-Thiet Nguyen
    Li, Sitan
    Cheah, Chien Chern
    IEEE ACCESS, 2022, 10 : 14270 - 14287
  • [3] Activation Distribution-based Layer-wise Quantization for Convolutional Neural Networks
    Ki, Subin
    Kim, Hyun
    2022 INTERNATIONAL CONFERENCE ON ELECTRONICS, INFORMATION, AND COMMUNICATION (ICEIC), 2022,
  • [4] Layer-Wise Weight Decay for Deep Neural Networks
    Ishii, Masato
    Sato, Atsushi
    IMAGE AND VIDEO TECHNOLOGY (PSIVT 2017), 2018, 10749 : 276 - 289
  • [5] A layer-wise Perturbation based Privacy Preserving Deep Neural Networks
    Adesuyi, Tosin A.
    Kim, Byeong Man
    2019 1ST INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION (ICAIIC 2019), 2019, : 389 - 394
  • [6] Stochastic Layer-Wise Precision in Deep Neural Networks
    Lacey, Griffin
    Taylor, Graham W.
    Areibi, Shawki
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2018, : 663 - 672
  • [7] Investigating Learning in Deep Neural Networks Using Layer-Wise Weight Change
    Agrawal, Ayush Manish
    Tendle, Atharva
    Sikka, Harshvardhan
    Singh, Sahib
    Kayid, Amr
    INTELLIGENT COMPUTING, VOL 2, 2021, 284 : 678 - 693
  • [8] Unsupervised Layer-Wise Model Selection in Deep Neural Networks
    Ludovic, Arnold
    Helene, Paugam-Moisy
    Michele, Sebag
    ECAI 2010 - 19TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2010, 215 : 915 - 920
  • [9] Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon
    Dong, Xin
    Chen, Shangyu
    Pan, Sinno Jialin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [10] Layer-wise contribution-filtered propagation for deep learning-based fault isolation
    Pan, Zhuofu
    Wang, Yalin
    Wang, Kai
    Ran, Guangtao
    Chen, Hongtian
    Gui, Weihua
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2022, 32 (17) : 9120 - 9138