Learning to Generate Diverse Data From a Temporal Perspective for Data-Free Quantization

被引:0
作者
Luo, Hui [1 ,2 ,3 ]
Zhang, Shuhai [4 ,5 ]
Zhuang, Zhuangwei [4 ,5 ]
Mai, Jiajie [4 ]
Tan, Mingkui [4 ,5 ]
Zhang, Jianlin [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Natl Key Lab Opt Field Manipulat Sci & Technol, Chengdu 610209, Peoples R China
[2] Chinese Acad Sci, Inst Opt & Elect, Key Lab Opt Engn, Chengdu 610209, Peoples R China
[3] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
[4] South China Univ Technol, Sch Software Engn, Guangzhou 510641, Peoples R China
[5] South China Univ Technol, Minist Educ, Key Lab Big Data & Intelligent Robot, Guangzhou 510641, Peoples R China
基金
中国国家自然科学基金;
关键词
Quantization (signal); Data models; Synthetic data; Generators; Computational modeling; Training; Analytical models; Model quantization; data-free quantization; generation process; synthetic data; linear interpolation; BINARY NEURAL-NETWORKS; ACCURATE;
D O I
10.1109/TCSVT.2024.3399311
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Model quantization is a prevalent method to compress and accelerate neural networks. Most existing quantization methods usually require access to real data to improve the performance of quantized models, which is often infeasible in some scenarios with privacy and security concerns. Recently, data-free quantization has been widely studied to solve the challenge of not having access to real data by generating synthetic data, among which generator-based data-free quantization is an important type. Previous generator-based methods focus on improving the performance of quantized models by optimizing the spatial distribution of synthetic data, while ignoring the study of changes in synthetic data from a temporal perspective. In this work, we reveal that generator-based data-free quantization methods usually suffer from the issue that synthetic data show homogeneity in the mid-to-late stages of the generation process due to the stagnation of the generator update, which hinders further improvement of the performance of quantized models. To solve the above issue, we propose introducing the discrepancy between the full-precision and quantized models as new supervision information to update the generator. Specifically, we propose a simple yet effective adversarial Gaussian-margin loss, which promotes continuous updating of the generator by adding more supervision information to the generator when the discrepancy between the full-precision and quantized models is small, thereby generating heterogeneous synthetic data. Moreover, to mitigate the homogeneity of the synthetic data further, we augment the synthetic data with linear interpolation. Our proposed method can also promote the performance of other generator-based data-free quantization methods. Extensive experimental results show that our proposed method achieves superior performances for various settings on data-free quantization, especially in ultra-low-bit settings, such as 3-bit.
引用
收藏
页码:9484 / 9498
页数:15
相关论文
共 50 条
  • [31] Accelerating Federated Learning for IoT in Big Data Analytics With Pruning, Quantization and Selective Updating
    Xu, Wenyuan
    Fang, Weiwei
    Ding, Yi
    Zou, Meixia
    Xiong, Naixue
    IEEE ACCESS, 2021, 9 : 38457 - 38466
  • [32] Over-the-Air Federated Learning From Heterogeneous Data
    Sery, Tomer
    Shlezinger, Nir
    Cohen, Kobi
    Eldar, Yonina C.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 3796 - 3811
  • [33] Data Quality-Aware Mixed-Precision Quantization via Hybrid Reinforcement Learning
    Wang, Yingchun
    Guo, Song
    Guo, Jingcai
    Zhang, Yuanhong
    Zhang, Weizhan
    Zheng, Qinghua
    Zhang, Jie
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 14
  • [34] Learning to Generate Parameters of ConvNets for Unseen Image Data
    Wang, Shiye
    Feng, Kaituo
    Li, Changsheng
    Yuan, Ye
    Wang, Guoren
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 5577 - 5592
  • [35] Adjusting Learning Depth in Nonnegative Latent Factorization of Tensors for Accurately Modeling Temporal Patterns in Dynamic QoS Data
    Luo, Xin
    Chen, Minzhi
    Wu, Hao
    Liu, Zhigang
    Yuan, Huaqiang
    Zhou, Mengchu
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2021, 18 (04) : 2142 - 2155
  • [36] CDFKD-MFS: Collaborative Data-Free Knowledge Distillation via Multi-Level Feature Sharing
    Hao, Zhiwei
    Luo, Yong
    Wang, Zhi
    Hu, Han
    An, Jianping
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 4262 - 4274
  • [37] Visual Analytics for Machine Learning: A Data Perspective Survey
    Wang, Junpeng
    Liu, Shixia
    Zhang, Wei
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (12) : 7637 - 7656
  • [38] CBFL: A Communication-Efficient Federated Learning Framework From Data Redundancy Perspective
    Li, Kaiju
    Xiao, Chunhua
    IEEE SYSTEMS JOURNAL, 2022, 16 (04): : 5572 - 5583
  • [39] Deep Learning for Spatio-Temporal Data Mining: A Survey
    Wang, Senzhang
    Cao, Jiannong
    Yu, Philip S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (08) : 3681 - 3700
  • [40] Improving Data Cleaning by Learning From Unstructured Textual Data
    Nasfi, Rihem
    de Tre, Guy
    Bronselaer, Antoon
    IEEE ACCESS, 2025, 13 : 36470 - 36491