Towards Feature Distribution Alignment and Diversity Enhancement for Data-Free Quantization

被引:6
作者
Gao, Yangcheng [1 ,2 ,3 ]
Zhang, Zhao [1 ,2 ,3 ]
Hong, Richang [1 ,2 ,3 ]
Zhang, Haijun [4 ]
Fan, Jicong [5 ,6 ]
Yan, Shuicheng [7 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
[2] Hefei Univ Technol, Minist Educ, Key Lab Knowledge Engn Big Data, Hefei 230009, Peoples R China
[3] Hefei Univ Technol, Intelligent Interconnected Syst Lab Anhui Prov, Hefei 230009, Peoples R China
[4] Harbin Inst Technol Shenzhen, Dept Comp Sci, Shenzhen, Peoples R China
[5] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen, Peoples R China
[6] Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
[7] Natl Univ Singapore, Singapore 117583, Singapore
来源
2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM) | 2022年
基金
中国国家自然科学基金;
关键词
Model compression; data-free low-bit model quantization; less performance loss; feature distribution alignment; diversity enhancement;
D O I
10.1109/ICDM54844.2022.00024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To obtain lower inference latency and less memory footprint of deep neural networks, model quantization has been widely employed in deep model deployment, by converting the floating points to low-precision integers. However, previous methods (such as quantization aware training and post training quantization) require original data for the fine-tuning or calibration of quantized model, which makes them inapplicable to the cases that original data are not accessed due to privacy or security. This gives birth to the data-free quantization method with synthetic data generation. While current data-free quantization methods still suffer from severe performance degradation when quantizing a model into lower bit, caused by the low inter-class separability of semantic features. To this end, we propose a new and effective data-free quantization method termed ClusterQ, which utilizes the feature distribution alignment for synthetic data generation. To obtain high inter-class separability of semantic features, we cluster and align the feature distribution statistics to imitate the distribution of real data, so that the performance degradation is alleviated. Moreover, we incorporate the diversity enhancement to solve class-wise mode collapse. We also employ the exponential moving average to update the centroid of each cluster for further feature distribution improvement. Extensive experiments based on different deep models (e.g., ResNet-18 and MobileNet-V2) over the ImageNet dataset demonstrate that our proposed ClusterQ model obtains state-of-the-art performance.
引用
收藏
页码:141 / 150
页数:10
相关论文
共 44 条
  • [41] Dual-Constrained Deep Semi-Supervised Coupled Factorization Network with Enriched Prior
    Zhang, Yan
    Zhang, Zhao
    Wang, Yang
    Zhang, Zheng
    Zhang, Li
    Yan, Shuicheng
    Wang, Meng
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (12) : 3233 - 3254
  • [42] Discriminative Additive Scale Loss for Deep Imbalanced Classification and Embedding
    Zhang, Zhao
    Jiang, Weiming
    Wang, Yang
    Ye, Qiaolin
    Zhao, Mingbo
    Xu, Mingliang
    Wang, Meng
    [J]. 2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 936 - 945
  • [43] Zhong YS, 2022, Arxiv, DOI arXiv:2109.04186
  • [44] Zhu B., 2021, arXiv