Generative Data Free Model Quantization With Knowledge Matching for Classification

被引:7
|
作者
Xu, Shoukai [1 ,2 ]
Zhang, Shuhai [1 ]
Liu, Jing [3 ]
Zhuang, Bohan [3 ]
Wang, Yaowei [2 ]
Tan, Mingkui [1 ,4 ]
机构
[1] South China Univ Technol, Sch Software Engn, Guangzhou 510006, Peoples R China
[2] Peng Cheng Lab, Shenzhen 518055, Peoples R China
[3] Monash Univ, Fac Informat Technol, Clayton, Vic 3800, Australia
[4] South China Univ Technol, Key Lab Big Data & Intelligent Robot, Minist Educ, Guangzhou 510006, Peoples R China
关键词
Data privacy and security; model compression; data free quantization; data generation; BINARY NEURAL-NETWORKS; IMAGE; SEGMENTATION; CONVOLUTION; ACCURATE; CNN;
D O I
10.1109/TCSVT.2023.3279281
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Neural network quantization aims to reduce the model size, computational complexity, and memory consumption by mapping weights and activations from full-precision to low-precision. However, many existing quantization methods, either post-training with calibration or quantization-aware training with fine-tuning, require original data for better performance, which may not be available due to confidentiality or privacy constraints. This lack of data can lead to a significant decline in performance. In this paper, we propose a universal and effective method called Generative Data Free Model Quantization with Knowledge Matching for Classification(KMDFQ) that removes the dependence on data for neural network quantization. To achieve this, we propose a knowledge matching generator that produces meaningful fake data based on the latent knowledge in the pre-trained model, including classification boundary knowledge and data distribution information. Based on this generator, we propose a fake-data driven data free quantization method that uses the generated data to take advantage of the latent knowledge for quantization. Furthermore, we introduce Mean Square Error alignment during the fine-tuning of the quantized model to more strictly and directly learn knowledge, making it more suitable for data free quantization. Extensive experiments on image classification demonstrate the effectiveness of our method, achieving higher accuracy than existing data free quantization methods, particularly as the quantization bit decreases. For example, on ImageNet, the 4-bit data free quantized ResNet-18 has less than a 1.2% accuracy decline compared to quantization with real data. The source code is available at https://github.com/ZSHsh98/KMDFQ.
引用
收藏
页码:7296 / 7309
页数:14
相关论文
共 50 条
  • [1] Diverse Sample Generation: Pushing the Limit of Generative Data-Free Quantization
    Qin, Haotong
    Ding, Yifu
    Zhang, Xiangguo
    Wang, Jiakai
    Liu, Xianglong
    Lu, Jiwen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 11689 - 11706
  • [2] Conditional generative data-free knowledge distillation
    Yu, Xinyi
    Yan, Ling
    Yang, Yang
    Zhou, Libo
    Ou, Linlin
    IMAGE AND VISION COMPUTING, 2023, 131
  • [3] Learning to Generate Diverse Data From a Temporal Perspective for Data-Free Quantization
    Luo, Hui
    Zhang, Shuhai
    Zhuang, Zhuangwei
    Mai, Jiajie
    Tan, Mingkui
    Zhang, Jianlin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (10) : 9484 - 9498
  • [4] Image Watermarking Based Data Hiding by Discrete Wavelet Transform Quantization Model with Convolutional Generative Adversarial Architectures
    Annadurai, C.
    Nelson, I.
    Devi, K. Nirmala
    Manikandan, R.
    Gandomi, Amir H.
    APPLIED SCIENCES-BASEL, 2023, 13 (02):
  • [5] Patch Similarity Aware Data-Free Quantization for Vision Transformers
    Li, Zhikai
    Ma, Liping
    Chen, Mengjuan
    Xiao, Junrui
    Gu, Qingyi
    COMPUTER VISION, ECCV 2022, PT XI, 2022, 13671 : 154 - 170
  • [6] Frequency Domain Distillation for Data-Free Quantization of Vision Transformer
    Nan, Gongrui
    Chao, Fei
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 205 - 216
  • [7] Dual-discriminator adversarial framework for data-free quantization
    Li, Zhikai
    Ma, Liping
    Long, Xianlei
    Xiao, Junrui
    Gu, Qingyi
    NEUROCOMPUTING, 2022, 511 : 67 - 77
  • [8] Adaptive knowledge transfer for data-free low-bit quantization via tiered collaborative
    Lin, Tong
    Li, Chenyang
    Qian, Bo
    Yang, Xinyu
    Wei, Xing
    Yang, Zelin
    NEUROCOMPUTING, 2025, 638
  • [9] Analyzing DDoS Attack Classification with Data Imbalance Using Generative Adversarial Networks
    Acosta-Tejada, Danny E.
    Sanchez-Galan, Javier E.
    Torres-Batista, Nelliud
    2023 IEEE LATIN-AMERICAN CONFERENCE ON COMMUNICATIONS, LATINCOM, 2023,
  • [10] Data-Free Quantization with Accurate Activation Clipping and Adaptive Batch Normalization
    He, Yefei
    Zhang, Luoming
    Wu, Weijia
    Zhou, Hong
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 10555 - 10568