AdaDFKD: Exploring adaptive inter-sample relationship in data-free knowledge distillation

被引:0
|
作者
Li, Jingru [1 ]
Zhou, Sheng [1 ]
Li, Liangcheng [1 ]
Wang, Haishuai [1 ]
Bu, Jiajun [1 ]
Yu, Zhi [1 ,2 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Zheda Rd, Hangzhou 310027, Zhejiang, Peoples R China
[2] Zhejiang Univ, Zhejiang Prov Key Lab Serv Robot, Zheda Rd, Hangzhou 310027, Zhejiang, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Data-free knowledge distillation; Unsupervised representation learning; Knowledge distillation;
D O I
10.1016/j.neunet.2024.106386
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In scenarios like privacy protection or large-scale data transmission, data-free knowledge distillation (DFKD) methods are proposed to learn Knowledge Distillation (KD) when data is not accessible. They generate pseudo samples by extracting the knowledge from teacher model, and utilize above pseudo samples for KD. The challenge in previous DFKD methods lies in the static nature of their target distributions and they focus on learning the instance-level distributions, causing its reliance on the pretrained teacher model. To address above concerns, our study introduces a novel DFKD approach known as AdaDFKD, designed to establish and utilize relationships among pseudo samples, which is adaptive to the student model, and finally effectively mitigates the aforementioned risk. We achieve this by generating from "easy-to-discriminate"samples to "hardto-discriminate"samples as human does. We design a relationship refinement module (R2M) to optimize the generation process, wherein we learn a progressive conditional distribution of negative samples and maximize the log-likelihood of inter-sample similarity of pseudosamples. Theoretically, we discover that such design of AdaDFKD both minimize the divergence and maximize the mutual information between the distribution of teacher and student models. Above results demonstrate the superiority of our approach over state-of-the-art (SOTA) DFKD methods across various benchmarks, teacher-student pairs, and evaluation metrics, as well as robustness and fast convergence.
引用
收藏
页数:15
相关论文
共 37 条
  • [11] Conditional pseudo-supervised contrast for data-Free knowledge distillation
    Shao, Renrong
    Zhang, Wei
    Wang, Jun
    PATTERN RECOGNITION, 2023, 143
  • [12] D3K: Dynastic Data-Free Knowledge Distillation
    Li, Xiufang
    Sun, Qigong
    Jiao, Licheng
    Liu, Fang
    Liu, Xu
    Li, Lingling
    Chen, Puhua
    Zuo, Yi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 8358 - 8371
  • [13] A Category-Aware Curriculum Learning for Data-Free Knowledge Distillation
    Li, Xiufang
    Jiao, Licheng
    Sun, Qigong
    Liu, Fang
    Liu, Xu
    Li, Lingling
    Chen, Puhua
    Yang, Shuyuan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 9603 - 9618
  • [14] Data-Free Knowledge Distillation for Privacy-Preserving Efficient UAV Networks
    Yu, Guyang
    2022 6TH INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION SCIENCES (ICRAS 2022), 2022, : 52 - 56
  • [15] Effective and efficient conditional contrast for data-free knowledge distillation with low memory
    Jiang, Chenyang
    Li, Zhendong
    Yang, Jun
    Wu, Yiqiang
    Li, Shuai
    JOURNAL OF SUPERCOMPUTING, 2025, 81 (04)
  • [16] Reusable generator data-free knowledge distillation with hard loss simulation for image classification
    Sun, Yafeng
    Wang, Xingwang
    Huang, Junhong
    Chen, Shilin
    Hou, Minghui
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [17] Data-Free Ensemble Knowledge Distillation for Privacy-conscious Multimedia Model Compression
    Hao, Zhiwei
    Luo, Yong
    Hu, Han
    An, Jianping
    Wen, Yonggang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1803 - 1811
  • [18] Memory efficient data-free distillation for continual learning
    Li, Xiaorong
    Wang, Shipeng
    Sun, Jian
    Xu, Zongben
    PATTERN RECOGNITION, 2023, 144
  • [19] Data-free knowledge distillation via generator-free data generation for Non-IID federated learning
    Zhao, Siran
    Liao, Tianchi
    Fu, Lele
    Chen, Chuan
    Bian, Jing
    Zheng, Zibin
    NEURAL NETWORKS, 2024, 179
  • [20] FedAlign: Federated Model Alignment via Data-Free Knowledge Distillation for Machine Fault Diagnosis
    Sun, Wenjun
    Yan, Ruqiang
    Jin, Ruibing
    Zhao, Rui
    Chen, Zhenghua
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 12