Towards Feature Distribution Alignment and Diversity Enhancement for Data-Free Quantization

被引:6
作者
Gao, Yangcheng [1 ,2 ,3 ]
Zhang, Zhao [1 ,2 ,3 ]
Hong, Richang [1 ,2 ,3 ]
Zhang, Haijun [4 ]
Fan, Jicong [5 ,6 ]
Yan, Shuicheng [7 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
[2] Hefei Univ Technol, Minist Educ, Key Lab Knowledge Engn Big Data, Hefei 230009, Peoples R China
[3] Hefei Univ Technol, Intelligent Interconnected Syst Lab Anhui Prov, Hefei 230009, Peoples R China
[4] Harbin Inst Technol Shenzhen, Dept Comp Sci, Shenzhen, Peoples R China
[5] Chinese Univ Hong Kong, Sch Data Sci, Shenzhen, Peoples R China
[6] Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
[7] Natl Univ Singapore, Singapore 117583, Singapore
来源
2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM) | 2022年
基金
中国国家自然科学基金;
关键词
Model compression; data-free low-bit model quantization; less performance loss; feature distribution alignment; diversity enhancement;
D O I
10.1109/ICDM54844.2022.00024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To obtain lower inference latency and less memory footprint of deep neural networks, model quantization has been widely employed in deep model deployment, by converting the floating points to low-precision integers. However, previous methods (such as quantization aware training and post training quantization) require original data for the fine-tuning or calibration of quantized model, which makes them inapplicable to the cases that original data are not accessed due to privacy or security. This gives birth to the data-free quantization method with synthetic data generation. While current data-free quantization methods still suffer from severe performance degradation when quantizing a model into lower bit, caused by the low inter-class separability of semantic features. To this end, we propose a new and effective data-free quantization method termed ClusterQ, which utilizes the feature distribution alignment for synthetic data generation. To obtain high inter-class separability of semantic features, we cluster and align the feature distribution statistics to imitate the distribution of real data, so that the performance degradation is alleviated. Moreover, we incorporate the diversity enhancement to solve class-wise mode collapse. We also employ the exponential moving average to update the centroid of each cluster for further feature distribution improvement. Extensive experiments based on different deep models (e.g., ResNet-18 and MobileNet-V2) over the ImageNet dataset demonstrate that our proposed ClusterQ model obtains state-of-the-art performance.
引用
收藏
页码:141 / 150
页数:10
相关论文
共 44 条
  • [31] Paszke A., 2017, AUTOMATIC DIFFERENTI
  • [32] Robust Low-rank Deep Feature Recovery in CNNs: Toward Low Information Loss and Fast Convergence
    Ren, Jiahuan
    Zhang, Zhao
    Fan, Jicong
    Zhang, Haijun
    Xu, Mingliang
    Wang, Meng
    [J]. 2021 21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2021), 2021, : 529 - 538
  • [33] EDP: An Efficient Decomposition and Pruning Scheme for Convolutional Neural Network Compression
    Ruan, Xiaofeng
    Liu, Yufan
    Yuan, Chunfeng
    Li, Bing
    Hu, Weiming
    Li, Yangxi
    Maybank, Stephen
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4499 - 4513
  • [34] MobileNetV2: Inverted Residuals and Linear Bottlenecks
    Sandler, Mark
    Howard, Andrew
    Zhu, Menglong
    Zhmoginov, Andrey
    Chen, Liang-Chieh
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 4510 - 4520
  • [35] Shoukai Xu, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12357), P1, DOI 10.1007/978-3-030-58610-2_1
  • [36] Szegedy C, 2015, PROC CVPR IEEE, P1, DOI 10.1109/CVPR.2015.7298594
  • [37] Rethinking Feature Distribution for Loss Functions in Image Classification
    Wan, Weitao
    Zhong, Yuanyi
    Li, Tianpeng
    Chen, Jiansheng
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 9117 - 9126
  • [38] Real-Time Correlation Tracking Via Joint Model Compression and Transfer
    Wang, Ning
    Zhou, Wengang
    Song, Yibing
    Ma, Chao
    Li, Houqiang
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 6123 - 6135
  • [39] DerainCycleGAN: Rain Attentive CycleGAN for Single Image Deraining and Rainmaking
    Wei, Yanyan
    Zhang, Zhao
    Wang, Yang
    Xu, Mingliang
    Yang, Yi
    Yan, Shuicheng
    Wang, Meng
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 4788 - 4801
  • [40] Diversifying Sample Generation for Accurate Data-Free Quantization
    Zhang, Xiangguo
    Qin, Haotong
    Ding, Yifu
    Gong, Ruihao
    Yan, Qinghua
    Tao, Renshuai
    Li, Yuhang
    Yu, Fengwei
    Liu, Xianglong
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15653 - 15662