Adaptive Data-Free Quantization

被引:27
|
作者
Qian, Biao [1 ]
Wang, Yang [1 ]
Hong, Richang [1 ]
Wang, Meng [1 ]
机构
[1] Hefei Univ Technol, Key Lab Knowledge Engn Big Data, Minist Educ, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR | 2023年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.00769
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data-free quantization (DFQ) recovers the performance of quantized network (Q) without the original data, but generates the fake sample via a generator (G) by learning from full-precision network (P), which, however, is totally independent of Q, overlooking the adaptability of the knowledge from generated samples, i.e., informative or not to the learning process of Q, resulting into the overflow of generalization error. Building on this, several critical questions how to measure the sample adaptability to Q under varied bit-width scenarios? whether the largest adaptability is the best? how to generate the samples with adaptive adaptability to improve Q's generalization? To answer the above questions, in this paper, we propose an Adaptive Data-Free Quantization (AdaDFQ) method, which revisits DFQ from a zero-sum game perspective upon the sample adaptability between two players - a generator and a quantized network. Following this viewpoint, we further define the disagreement and agreement samples to form two boundaries, where the margin between two boundaries is optimized to adaptively regulate the adaptability of generated samples to Q, so as to address the over-and-under fitting issues. Our AdaDFQ reveals: 1) the largest adaptability is NOT the best for sample generation to benefit Q's generalization; 2) the knowledge of the generated sample should not be informative to Q only, but also related to the category and distribution information of the training data for P. The theoretical and empirical analysis validate the advantages of AdaDFQ over the state-of-the-arts. Our code is available at https://github.com/hfutqian/AdaDFQ.
引用
收藏
页码:7960 / 7968
页数:9
相关论文
共 50 条
  • [1] Data-Free Quantization with Accurate Activation Clipping and Adaptive Batch Normalization
    He, Yefei
    Zhang, Luoming
    Wu, Weijia
    Zhou, Hong
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 10555 - 10568
  • [2] Data-Free Quantization with Accurate Activation Clipping and Adaptive Batch Normalization
    Yefei He
    Luoming Zhang
    Weijia Wu
    Hong Zhou
    Neural Processing Letters, 2023, 55 : 10555 - 10568
  • [3] Data-Free Network Quantization With Adversarial Knowledge Distillation
    Choi, Yoojin
    Choi, Jihwan
    El-Khamy, Mostafa
    Lee, Jungwon
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 3047 - 3057
  • [4] META-BNS FOR ADVERSARIAL DATA-FREE QUANTIZATION
    Fu, Siming
    Wang, Hualiang
    Cao, Yuchen
    Hu, Haoji
    Peng, Bo
    Tan, Wenming
    Ye, Tingqun
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 4038 - 4042
  • [5] Diversifying Sample Generation for Accurate Data-Free Quantization
    Zhang, Xiangguo
    Qin, Haotong
    Ding, Yifu
    Gong, Ruihao
    Yan, Qinghua
    Tao, Renshuai
    Li, Yuhang
    Yu, Fengwei
    Liu, Xianglong
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15653 - 15662
  • [6] EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs
    Tang, Hanlin
    Sun, Yifu
    Wu, Decheng
    Liu, Kai
    Zhu, Jianchen
    Kang, Zhanhui
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9119 - 9128
  • [7] Data-Free Quantization ThroughWeight Equalization and Bias Correction
    Nagel, Markus
    van Baalen, Mart
    Blankevoort, Tijmen
    Welling, Max
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 1325 - 1334
  • [8] REx: Data-Free Residual Quantization Error Expansion
    Yvinec, Edouard
    Dapogny, Arnaud
    Cord, Matthieu
    Bailly, Kevin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Adaptive knowledge transfer for data-free low-bit quantization via tiered collaborative
    Lin, Tong
    Li, Chenyang
    Qian, Bo
    Yang, Xinyu
    Wei, Xing
    Yang, Zelin
    NEUROCOMPUTING, 2025, 638
  • [10] Rethinking Data-Free Quantization as a Zero-Sum Game
    Qian, Biao
    Wang, Yang
    Hong, Richang
    Wang, Meng
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9489 - 9497