Learning to Generate Diverse Data From a Temporal Perspective for Data-Free Quantization

被引:0
|
作者
Luo, Hui [1 ,2 ,3 ]
Zhang, Shuhai [4 ,5 ]
Zhuang, Zhuangwei [4 ,5 ]
Mai, Jiajie [4 ]
Tan, Mingkui [4 ,5 ]
Zhang, Jianlin [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Natl Key Lab Opt Field Manipulat Sci & Technol, Chengdu 610209, Peoples R China
[2] Chinese Acad Sci, Inst Opt & Elect, Key Lab Opt Engn, Chengdu 610209, Peoples R China
[3] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
[4] South China Univ Technol, Sch Software Engn, Guangzhou 510641, Peoples R China
[5] South China Univ Technol, Minist Educ, Key Lab Big Data & Intelligent Robot, Guangzhou 510641, Peoples R China
基金
中国国家自然科学基金;
关键词
Quantization (signal); Data models; Synthetic data; Generators; Computational modeling; Training; Analytical models; Model quantization; data-free quantization; generation process; synthetic data; linear interpolation; BINARY NEURAL-NETWORKS; ACCURATE;
D O I
10.1109/TCSVT.2024.3399311
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Model quantization is a prevalent method to compress and accelerate neural networks. Most existing quantization methods usually require access to real data to improve the performance of quantized models, which is often infeasible in some scenarios with privacy and security concerns. Recently, data-free quantization has been widely studied to solve the challenge of not having access to real data by generating synthetic data, among which generator-based data-free quantization is an important type. Previous generator-based methods focus on improving the performance of quantized models by optimizing the spatial distribution of synthetic data, while ignoring the study of changes in synthetic data from a temporal perspective. In this work, we reveal that generator-based data-free quantization methods usually suffer from the issue that synthetic data show homogeneity in the mid-to-late stages of the generation process due to the stagnation of the generator update, which hinders further improvement of the performance of quantized models. To solve the above issue, we propose introducing the discrepancy between the full-precision and quantized models as new supervision information to update the generator. Specifically, we propose a simple yet effective adversarial Gaussian-margin loss, which promotes continuous updating of the generator by adding more supervision information to the generator when the discrepancy between the full-precision and quantized models is small, thereby generating heterogeneous synthetic data. Moreover, to mitigate the homogeneity of the synthetic data further, we augment the synthetic data with linear interpolation. Our proposed method can also promote the performance of other generator-based data-free quantization methods. Extensive experimental results show that our proposed method achieves superior performances for various settings on data-free quantization, especially in ultra-low-bit settings, such as 3-bit.
引用
收藏
页码:9484 / 9498
页数:15
相关论文
共 50 条
  • [21] Latent Code Augmentation Based on Stable Diffusion for Data-Free Substitute Attacks
    Shao, Mingwen
    Meng, Lingzhuang
    Qiao, Yuanjian
    Zhang, Lixu
    Zuo, Wangmeng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025,
  • [22] Data-Free Low-Bit Quantization via Dynamic Multi-teacher Knowledge Distillation
    Huang, Chong
    Lin, Shaohui
    Zhang, Yan
    Li, Ke
    Zhang, Baochang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 28 - 41
  • [23] Spatio-Temporal Prediction on Streaming Data: A Unified Federated Continuous Learning Framework
    Miao, Hao
    Zhao, Yan
    Guo, Chenjuan
    Yang, Bin
    Zheng, Kai
    Jensen, Christian S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (04) : 2126 - 2140
  • [24] Generative Data Free Model Quantization With Knowledge Matching for Classification
    Xu, Shoukai
    Zhang, Shuhai
    Liu, Jing
    Zhuang, Bohan
    Wang, Yaowei
    Tan, Mingkui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7296 - 7309
  • [25] PSAQ-ViT V2: Toward Accurate and General Data-Free Quantization for Vision Transformers
    Li, Zhikai
    Chen, Mengjuan
    Xiao, Junrui
    Gu, Qingyi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 17227 - 17238
  • [26] Data-Free Solution of Electromagnetic PDEs Using Neural Networks and Extension to Transfer Learning
    Bhardwaj, Shubhendu
    Gaire, Pawan
    IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, 2022, 70 (07) : 5179 - 5188
  • [27] VDM-DA: Virtual Domain Modeling for Source Data-Free Domain Adaptation
    Tian, Jiayi
    Zhang, Jing
    Li, Wen
    Xu, Dong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (06) : 3749 - 3760
  • [28] A flexible data-free framework for structure-based de novo drug design with reinforcement learning
    Du, Hongyan
    Jiang, Dejun
    Zhang, Odin
    Wu, Zhenxing
    Gao, Junbo
    Zhang, Xujun
    Wang, Xiaorui
    Deng, Yafeng
    Kang, Yu
    Li, Dan
    Pan, Peichen
    Hsieh, Chang-Yu
    Hou, Tingjun
    CHEMICAL SCIENCE, 2023, 14 (43) : 12166 - 12181
  • [29] Neural-Sim: Learning to Generate Training Data with NeRF
    Ge, Yunhao
    Behl, Harkirat
    Xu, Jiashu
    Gunasekar, Suriya
    Joshi, Neel
    Song, Yale
    Wang, Xin
    Itti, Laurent
    Vineet, Vibhav
    COMPUTER VISION, ECCV 2022, PT XXIII, 2022, 13683 : 477 - 493
  • [30] FedAlign: Federated Model Alignment via Data-Free Knowledge Distillation for Machine Fault Diagnosis
    Sun, Wenjun
    Yan, Ruqiang
    Jin, Ruibing
    Zhao, Rui
    Chen, Zhenghua
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 12