A NOVEL SENSITIVITY METRIC FOR MIXED-PRECISION QUANTIZATION WITH SYNTHETIC DATA GENERATION

被引:1
|
作者
Lee, Donghyun [1 ]
Cho, Minkyoung [1 ]
Lee, Seungwon [1 ]
Song, Joonho [1 ]
Choi, Changkyu [1 ]
机构
[1] Samsung Elect, Samsung Adv Inst Technol, Suwon, South Korea
关键词
Deep Learning; Quantization; Data Free;
D O I
10.1109/ICIP42928.2021.9506527
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Post-training quantization is a representative technique for compressing neural networks, making them smaller and more efficient for deployment on edge devices. However, an inaccessible user dataset often makes it difficult to ensure the quality of the quantized neural network in practice. In addition, existing approaches may use a single uniform bit-width across the network, resulting in significant accuracy degradation at extremely low bit-widths. To utilize multiple bit-width, sensitivity metric plays a key role in balancing accuracy and compression. In this paper, we propose a novel sensitivity metric that considers the effect of quantization error on task loss and interaction with other layers. Moreover, we develop labeled data generation methods that are not dependent on a specific operation of the neural network. Our experiments show that the proposed metric better represents quantization sensitivity, and generated data are more feasible to apply to mixed-precision quantization.
引用
收藏
页码:1294 / 1298
页数:5
相关论文
共 50 条
  • [1] A Novel Mixed-Precision Quantization Approach for CNNs
    Wu, Dan
    Wang, Yanzhi
    Fei, Yuqi
    Gao, Guowang
    IEEE ACCESS, 2025, 13 : 49309 - 49319
  • [2] Scalable Data Generation for Evaluating Mixed-Precision Solvers
    Luszczek, Piotr
    Tsai, Yaohung
    Lindquist, Neil
    Anzt, Hartwig
    Dongarra, Jack
    2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,
  • [3] Automatic Mixed-Precision Quantization Search of BERT
    Zhao, Changsheng
    Hua, Ting
    Shen, Yilin
    Lou, Qian
    Jin, Hongxia
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3427 - 3433
  • [4] EVOLUTIONARY QUANTIZATION OF NEURAL NETWORKS WITH MIXED-PRECISION
    Liu, Zhenhua
    Zhang, Xinfeng
    Wang, Shanshe
    Ma, Siwei
    Gao, Wen
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2785 - 2789
  • [5] Hardware-Centric AutoML for Mixed-Precision Quantization
    Wang, Kuan
    Liu, Zhijian
    Lin, Yujun
    Lin, Ji
    Han, Song
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (8-9) : 2035 - 2048
  • [6] Mixed-Precision Collaborative Quantization for Fast Object Tracking
    Xie, Yefan
    Guo, Yanwei
    Hou, Xuan
    Zheng, Jiangbin
    ADVANCES IN BRAIN INSPIRED COGNITIVE SYSTEMS, BICS 2023, 2024, 14374 : 229 - 238
  • [7] One-Shot Model for Mixed-Precision Quantization
    Koryakovskiy, Ivan
    Yakovleva, Alexandra
    Buchnev, Valentin
    Isaev, Temur
    Odinokikh, Gleb
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 7939 - 7949
  • [8] CSMPQ: Class Separability Based Mixed-Precision Quantization
    Wang, Mingkai
    Jin, Taisong
    Zhang, Miaohui
    Yu, Zhengtao
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT I, 2023, 14086 : 544 - 555
  • [9] Hardware-Centric AutoML for Mixed-Precision Quantization
    Kuan Wang
    Zhijian Liu
    Yujun Lin
    Ji Lin
    Song Han
    International Journal of Computer Vision, 2020, 128 : 2035 - 2048
  • [10] AMED: Automatic Mixed-Precision Quantization for Edge Devices
    Kimhi, Moshe
    Rozen, Tal
    Mendelson, Avi
    Baskin, Chaim
    MATHEMATICS, 2024, 12 (12)