Towards Feature Distribution Alignment and Diversity Enhancement for Data-Free Quantization

被引:6
作者
Gao, Yangcheng [1 ,2 ,3 ]
Zhang, Zhao [1 ,2 ,3 ]
Hong, Richang [1 ,2 ,3 ]
Zhang, Haijun [4 ]
Fan, Jicong [5 ,6 ]
Yan, Shuicheng [7 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
[2] Hefei Univ Technol, Minist Educ, Key Lab Knowledge Engn Big Data, Hefei 230009, Peoples R China
[3] Hefei Univ Technol, Intelligent Interconnected Syst Lab Anhui Prov, Hefei 230009, Peoples R China
[4] Harbin Inst Technol Shenzhen, Dept Comp Sci, Shenzhen, Peoples R China
[5] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen, Peoples R China
[6] Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
[7] Natl Univ Singapore, Singapore 117583, Singapore
来源
2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM) | 2022年
基金
中国国家自然科学基金;
关键词
Model compression; data-free low-bit model quantization; less performance loss; feature distribution alignment; diversity enhancement;
D O I
10.1109/ICDM54844.2022.00024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To obtain lower inference latency and less memory footprint of deep neural networks, model quantization has been widely employed in deep model deployment, by converting the floating points to low-precision integers. However, previous methods (such as quantization aware training and post training quantization) require original data for the fine-tuning or calibration of quantized model, which makes them inapplicable to the cases that original data are not accessed due to privacy or security. This gives birth to the data-free quantization method with synthetic data generation. While current data-free quantization methods still suffer from severe performance degradation when quantizing a model into lower bit, caused by the low inter-class separability of semantic features. To this end, we propose a new and effective data-free quantization method termed ClusterQ, which utilizes the feature distribution alignment for synthetic data generation. To obtain high inter-class separability of semantic features, we cluster and align the feature distribution statistics to imitate the distribution of real data, so that the performance degradation is alleviated. Moreover, we incorporate the diversity enhancement to solve class-wise mode collapse. We also employ the exponential moving average to update the centroid of each cluster for further feature distribution improvement. Extensive experiments based on different deep models (e.g., ResNet-18 and MobileNet-V2) over the ImageNet dataset demonstrate that our proposed ClusterQ model obtains state-of-the-art performance.
引用
收藏
页码:141 / 150
页数:10
相关论文
共 44 条
  • [1] Banner R, 2019, ADV NEUR IN, V32
  • [2] Cai YH, 2020, PROC CVPR IEEE, P13166, DOI 10.1109/CVPR42600.2020.01318
  • [3] Dynamical Channel Pruning by Conditional Accuracy Change for Deep Neural Networks
    Chen, Zhiqiang
    Xu, Ting-Bing
    Du, Changde
    Liu, Cheng-Lin
    He, Huiguang
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (02) : 799 - 813
  • [4] Cheng X., 2020, P IEEECVF C COMPUTER, p12 925
  • [5] Choi J, 2018, Arxiv, DOI arXiv:1805.06085
  • [6] Choi K, 2021, ADV NEUR IN, V34
  • [7] Courbariaux M., 2015, ICLR
  • [8] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [9] Denton E, 2014, ADV NEUR IN, V27
  • [10] Esser S.K., 2020, INT C LEARN REPR