One Teacher is Enough: A Server-Clueless Federated Learning With Knowledge Distillation

被引:0
作者
Ning, Wanyi [1 ,2 ]
Qi, Qi [1 ,2 ]
Wang, Jingyu [1 ,2 ]
Zhu, Mengde [1 ,2 ]
Li, Shaolong [1 ,2 ]
Yang, Guang [3 ]
Liao, Jianxin [1 ,2 ]
机构
[1] Beijing Univ Posts & Telecommun, Sate Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[2] E Byte COM, Beijing 100191, Peoples R China
[3] Alibaba DAMO Acad, XG Lab, Beijing 100102, Peoples R China
基金
中国国家自然科学基金;
关键词
Federated learning; Privacy; knowledge distillation; privacy protection;
D O I
10.1109/TSC.2024.3414372
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine learning-based services offer intelligent solutions with powerful models. To enhance model robustness, Federated Learning (FL) emerges as a promising collaborative learning paradigm, which iteratively trains a global model through parameter exchange among multiple clients based on their local data. Generally, the local data are heterogeneous, which slows down convergence. Knowledge distillation is an effective technique against data heterogeneity while existing works distill the ensemble knowledge from local models, ignoring the natural global knowledge from the aggregated model. This places limitations on their algorithms, such as the need for proxy data or the necessary exposure of local models to the server, which is prohibited in most privacy-preserving FL with a clueless server. In this work, we propose FedDGT, a novel knowledge distillation method for industrial server-clueless FL. FedDGT regards the aggregated model as the only one teacher to impart its global knowledge into a generator and then regularizes the drifted local models through the generator, overcoming previous limitations and providing better privacy and scalability support. Extensive experiments demonstrate that FedDGT can achieve highly-competitive model performance while greatly reducing the communication rounds in a server-clueless scenario.
引用
收藏
页码:2704 / 2718
页数:15
相关论文
共 48 条
  • [11] Privacy-Preserving and Reliable Decentralized Federated Learning
    Gao, Yuanyuan
    Zhang, Lei
    Wang, Lulu
    Choo, Kim-Kwang Raymond
    Zhang, Rui
    [J]. IEEE TRANSACTIONS ON SERVICES COMPUTING, 2023, 16 (04) : 2879 - 2891
  • [12] Database Workload Capacity Planning using Time Series Analysis and Machine Learning
    Higginson, Antony S.
    Dediu, Mihaela
    Arsene, Octavian
    Paton, Norman W.
    Embury, Suzanne M.
    [J]. SIGMOD'20: PROCEEDINGS OF THE 2020 ACM SIGMOD INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2020, : 769 - 783
  • [13] Hinton G, 2015, Arxiv, DOI arXiv:1503.02531
  • [14] Hsieh K, 2017, PROCEEDINGS OF NSDI '17: 14TH USENIX SYMPOSIUM ON NETWORKED SYSTEMS DESIGN AND IMPLEMENTATION, P629
  • [15] Huang YT, 2021, AAAI CONF ARTIF INTE, V35, P7865
  • [16] Jeong E, 2023, Arxiv, DOI arXiv:1811.11479
  • [17] Karimireddy SP, 2020, PR MACH LEARN RES, V119
  • [18] Krizhevsky A., 2009, LEARNING MULTIPLE LA
  • [19] ImageNet Classification with Deep Convolutional Neural Networks
    Krizhevsky, Alex
    Sutskever, Ilya
    Hinton, Geoffrey E.
    [J]. COMMUNICATIONS OF THE ACM, 2017, 60 (06) : 84 - 90
  • [20] Gradient-based learning applied to document recognition
    Lecun, Y
    Bottou, L
    Bengio, Y
    Haffner, P
    [J]. PROCEEDINGS OF THE IEEE, 1998, 86 (11) : 2278 - 2324