Relation-Guided Adversarial Learning for Data-Free Knowledge Transfer

被引:0
作者
Liang, Yingping [1 ]
Fu, Ying [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Knowledge distillation; Data-free distillation; Transfer learning; Model quantization; Incremental learning; Generative model;
D O I
10.1007/s11263-024-02303-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data-free knowledge distillation transfers knowledge by recovering training data from a pre-trained model. Despite the recent success of seeking global data diversity, the diversity within each class and the similarity among different classes are largely overlooked, resulting in data homogeneity and limited performance. In this paper, we introduce a novel Relation-Guided Adversarial Learning method with triplet losses, which solves the homogeneity problem from two aspects. To be specific, our method aims to promote both intra-class diversity and inter-class confusion of the generated samples. To this end, we design two phases, an image synthesis phase and a student training phase. In the image synthesis phase, we construct an optimization process to push away samples with the same labels and pull close samples with different labels, leading to intra-class diversity and inter-class confusion, respectively. Then, in the student training phase, we perform an opposite optimization, which adversarially attempts to reduce the distance of samples of the same classes and enlarge the distance of samples of different classes. To mitigate the conflict of seeking high global diversity and keeping inter-class confusing, we propose a focal weighted sampling strategy by selecting the negative in the triplets unevenly within a finite range of distance. RGAL shows significant improvement over previous state-of-the-art methods in accuracy and data efficiency. Besides, RGAL can be inserted into state-of-the-art methods on various data-free knowledge transfer applications. Experiments on various benchmarks demonstrate the effectiveness and generalizability of our proposed method on various tasks, specially data-free knowledge distillation, data-free quantization, and non-exemplar incremental learning. Our code will be publicly available to the community.
引用
收藏
页码:2868 / 2885
页数:18
相关论文
共 50 条
[21]   CDFKD-MFS: Collaborative Data-Free Knowledge Distillation via Multi-Level Feature Sharing [J].
Hao, Zhiwei ;
Luo, Yong ;
Wang, Zhi ;
Hu, Han ;
An, Jianping .
IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 :4262-4274
[22]   Privacy-Preserving Student Learning with Differentially Private Data-Free Distillation [J].
Liu, Bochao ;
Lu, Jianghu ;
Wang, Pengju ;
Zhang, Junjie ;
Zeng, Dan ;
Qian, Zhenxing ;
Ge, Shiming .
2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
[23]   AdaDFKD: Exploring adaptive inter-sample relationship in data-free knowledge distillation [J].
Li, Jingru ;
Zhou, Sheng ;
Li, Liangcheng ;
Wang, Haishuai ;
Bu, Jiajun ;
Yu, Zhi .
NEURAL NETWORKS, 2024, 177
[24]   Data-Free Ensemble Knowledge Distillation for Privacy-conscious Multimedia Model Compression [J].
Hao, Zhiwei ;
Luo, Yong ;
Hu, Han ;
An, Jianping ;
Wen, Yonggang .
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, :1803-1811
[25]   Reusable generator data-free knowledge distillation with hard loss simulation for image classification [J].
Sun, Yafeng ;
Wang, Xingwang ;
Huang, Junhong ;
Chen, Shilin ;
Hou, Minghui .
EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
[26]   IFHE: Intermediate-Feature Heterogeneity Enhancement for Image Synthesis in Data-Free Knowledge Distillation [J].
Chen, Yi ;
Liu, Ning ;
Ren, Ao ;
Yang, Tao ;
Liu, Duo .
2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
[27]   Data-Free Low-Bit Quantization via Dynamic Multi-teacher Knowledge Distillation [J].
Huang, Chong ;
Lin, Shaohui ;
Zhang, Yan ;
Li, Ke ;
Zhang, Baochang .
PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 :28-41
[28]   The DNN learning method for few training data via knowledge transfer [J].
Nigaki Y. ;
Inoue K. ;
Yoshioka M. .
Inoue, Katsufumi (inoue@cs.osakafu-u.ac.jp), 1600, Institute of Electrical Engineers of Japan (140) :664-672
[29]   Enhancing Global Model Performance in Federated Learning With Non-IID Data Using a Data-Free Generative Diffusion Model [J].
Najafi, Mohammadreza ;
Daneshtalab, Masoud ;
Lee, Jeong-A ;
Saadloonia, Ghazal ;
Shin, Seokjoo .
IEEE ACCESS, 2024, 12 :148230-148239
[30]   MDTGAN: Multi domain generative adversarial transfer learning network for traffic data imputation [J].
Fang, Jie ;
He, Hangyu ;
Xu, Mengyun ;
Chen, Hongting .
EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255