A Personalized Federated Learning Method Based on Clustering and Knowledge Distillation

被引:3
作者
Zhang, Jianfei [1 ]
Shi, Yongqiang [1 ]
机构
[1] Changchun Univ Sci & Technol, Sch Comp Sci & Technol, Changchun 130031, Peoples R China
关键词
federated learning; knowledge distillation; clustering; data heterogeneity;
D O I
10.3390/electronics13050857
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) is a distributed machine learning paradigm under privacy preservation. However, data heterogeneity among clients leads to the shared global model obtained after training, which cannot fit the distribution of each client's dataset, and the performance of the model degrades. To address this problem, we proposed a personalized federated learning method based on clustering and knowledge distillation, called pFedCK. In this algorithm, each client has an interactive model that participates in global training and a personalized model that is only trained locally. Both of the models perform knowledge distillation with each other through the feature representation of the middle layer and the soft prediction of the model. In addition, in order to make an interaction model only obtaining the model information from the client, which has similar data distribution and avoids the interference of other heterogeneous information, the server will cluster the clients according to the similarity of the amount of parameter variation uploaded by different interaction models during every training round. By clustering clients, interaction models with similar data distributions can cooperate with each other to better fit the local dataset distribution. Thereby, the performance of personalized model can be improved by obtaining more valuable information indirectly. Finally, we conduct simulation experiments on three benchmark datasets under different data heterogeneity scenarios. Compared to the single model algorithms, the accuracy of pFedCK improved by an average of 23.4% and 23.8% over FedAvg and FedProx, respectively; compared to typical personalization algorithms, the accuracy of pFedCK improved by an average of 0.8% and 1.3%, and a maximum of 1.0% and 2.9% over FedDistill and FML.
引用
收藏
页数:15
相关论文
共 22 条
  • [1] Communication-Efficient and Model-Heterogeneous Personalized Federated Learning via Clustered Knowledge Transfer
    Cho, Yae Jee
    Wang, Jianyu
    Chirvolu, Tarun
    Joshi, Gauri
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2023, 17 (01) : 234 - 247
  • [2] Collins L, 2023, Arxiv, DOI [arXiv:2102.07078, DOI 10.48550/ARXIV.2102.07078]
  • [3] Deng YY, 2020, Arxiv, DOI arXiv:2003.13461
  • [4] FedDC: Federated Learning with Non-IID Data via Local Drift Decoupling and Correction
    Gao, Liang
    Fu, Huazhu
    Li, Li
    Chen, Yingwen
    Xu, Ming
    Xu, Cheng-Zhong
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 10102 - 10111
  • [5] An Efficient Framework for Clustered Federated Learning
    Ghosh, Avishek
    Chung, Jichan
    Yin, Dong
    Ramchandran, Kannan
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2022, 68 (12) : 8076 - 8091
  • [6] Hinton G, 2015, Arxiv, DOI arXiv:1503.02531
  • [7] Federated Learning With Taskonomy for Non-IID Data
    Jamali-Rad, Hadi
    Abdizadeh, Mohammad
    Singh, Anuj
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8719 - 8730
  • [8] Jeong E, 2023, Arxiv, DOI arXiv:1811.11479
  • [9] Krizhevsky A., 2009, Tech. Rep. 0
  • [10] Gradient-based learning applied to document recognition
    Lecun, Y
    Bottou, L
    Bengio, Y
    Haffner, P
    [J]. PROCEEDINGS OF THE IEEE, 1998, 86 (11) : 2278 - 2324