Conditional pseudo-supervised contrast for data-Free knowledge distillation

被引:7
作者
Shao, Renrong [1 ]
Zhang, Wei [1 ]
Wang, Jun [1 ]
机构
[1] East China Normal Univ, Sch Comp Sci & Technol, 3663 North Zhongshan Rd, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Model compression; Knowledge distillation; Representation learning; Contrastive learning; Privacy protection;
D O I
10.1016/j.patcog.2023.109781
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data-free knowledge distillation (DFKD) is an effective manner to solve model compression and trans-mission restrictions while retaining privacy protection, which has attracted extensive attention in recent years. Currently, the majority of existing methods utilize a generator to synthesize images to support the distillation. Although the current methods have achieved great success, there are still many issues to be explored. Firstly, the outstanding performance of supervised learning in deep learning drives us to explore a pseudo-supervised paradigm on DFKD. Secondly, current synthesized methods cannot distin-guish the distributions of different categories of samples, thus producing ambiguous samples that may lead to an incorrect evaluation by the teacher. Besides, current methods cannot optimize the category -wise diversity samples, which will hinder the student model learning from diverse samples and further achieving better performance. In this paper, to address the above limitations, we propose a novel learning paradigm, i.e., conditional pseudo-supervised contrast for data-free knowledge distillation (CPSC-DFKD). The primary innovations of CPSC-DFKD are: (1) introducing a conditional generative adversarial network to synthesize category-specific diverse images for pseudo-supervised learning, (2) improving the mod-ules of the generator to distinguish the distributions of different categories, and (3) proposing pseudo -supervised contrastive learning based on teacher and student views to enhance diversity. Comprehensive experiments on three commonly-used datasets validate the performance lift of both the student and gen-erator brought by CPSC-DFKD. The code is available at https://github.com/RoryShao/CPSC-DFKD.git & COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 56 条
[41]   Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks [J].
Ren, Shaoqing ;
He, Kaiming ;
Girshick, Ross ;
Sun, Jian .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) :1137-1149
[42]   MobileNetV2: Inverted Residuals and Linear Bottlenecks [J].
Sandler, Mark ;
Howard, Andrew ;
Zhu, Menglong ;
Zhmoginov, Andrey ;
Chen, Liang-Chieh .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4510-4520
[43]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[44]   Class-specific discriminative metric learning for scene recognition [J].
Wang, Chen ;
Peng, Guohua ;
De Baets, Bernard .
PATTERN RECOGNITION, 2022, 126
[45]   Face Aging with Identity-Preserved Conditional Generative Adversarial Networks [J].
Wang, Zongwei ;
Tang, Xu ;
Luo, Weixin ;
Gao, Shenghua .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7939-7947
[46]   Generative Low-Bitwidth Data Free Quantization [J].
Xu, Shoukai ;
Li, Haokun ;
Zhuang, Bohan ;
Liu, Jing ;
Cao, Jiezhang ;
Liang, Chuangrun ;
Tan, Mingkui .
COMPUTER VISION - ECCV 2020, PT XII, 2020, 12357 :1-17
[47]   SAR-to-optical image translation based on improved CGAN [J].
Yang, Xi ;
Zhao, Jingyi ;
Wei, Ziyu ;
Wang, Nannan ;
Gao, Xinbo .
PATTERN RECOGNITION, 2022, 121
[48]   Deep neural network compression through interpretability-based filter pruning [J].
Yao, Kaixuan ;
Cao, Feilong ;
Leung, Yee ;
Liang, Jiye .
PATTERN RECOGNITION, 2021, 119
[49]   Dreaming to Distill: Data-free Knowledge Transfer via DeepInversion [J].
Yin, Hongxu ;
Molchanov, Pavlo ;
Alvarez, Jose M. ;
Li, Zhizhong ;
Mallya, Arun ;
Hoiem, Derek ;
Jha, Niraj K. ;
Kautz, Jan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :8712-8721
[50]   Data-Free Network Quantization With Adversarial Knowledge Distillation [J].
Choi, Yoojin ;
Choi, Jihwan ;
El-Khamy, Mostafa ;
Lee, Jungwon .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :3047-3057