Communication-Efficient and Model-Heterogeneous Personalized Federated Learning via Clustered Knowledge Transfer

被引:24
作者
Cho, Yae Jee [1 ]
Wang, Jianyu [1 ]
Chirvolu, Tarun [2 ]
Joshi, Gauri [1 ]
机构
[1] Carnegie Mellon Univ, Dept Elect & Comp Engn, Pittsburgh, PA 15213 USA
[2] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
关键词
Federated learning; communication efficiency; model heterogeneity; knowledge transfer; clustering;
D O I
10.1109/JSTSP.2022.3231527
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Personalized federated learning (PFL) aims to train model(s) that can perform well on the individual edge-devices' data where the edge-devices (clients) are usually IoT devices like our mobile phones. The participating clients for cross-device settings, in general, have heterogeneous system capabilities and limited communication bandwidth. Such practical properties of the edge-devices, however, are overlooked by many recent work in PFL, which use the same model architecture across all clients and incur high communication cost by directly communicating the model parameters. In our work, we propose a novel and practical PFL framework named COMET where clients can use heterogeneous models of their own choice and do not directly communicate their model parameters to other parties. Instead, COMET uses clustered codistillation, where clients use knowledge distillation to transfer their knowledge to other clients with similar data distributions. This presents a practical PFL framework for the edge-devices to train through IoT networks by lifting the heavy communication burden of communicating large models. We theoretically show the convergence and generalization properties of COMET and empirically show that COMET achieves high test accuracy with several orders of magnitude lower communication cost while allowing client model heterogeneity compared to the other state-of-the-art PFL methods.
引用
收藏
页码:234 / 247
页数:14
相关论文
共 50 条
  • [31] Lin X., 2022, PROC 39 INT C MACH L, V162, p13 344
  • [32] Aberrant Static and Dynamic Functional Network Connectivity in Acute Mild Traumatic Brain Injury with Cognitive Impairment
    Lu, Liyan
    Zhang, Juan
    Li, Fengfang
    Shang, Song'an
    Chen, Huiyou
    Yin, Xindao
    Gao, Wei
    Chen, Yu-Chen
    [J]. CLINICAL NEURORADIOLOGY, 2022, 32 (01) : 205 - 214
  • [33] Privacy and Robustness in Federated Learning: Attacks and Defenses
    Lyu, Lingjuan
    Yu, Han
    Ma, Xingjun
    Chen, Chen
    Sun, Lichao
    Zhao, Jun
    Yang, Qiang
    Yu, Philip S.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 8726 - 8746
  • [34] Adaptive Distillation for Decentralized Learning from Heterogeneous Clients
    Ma, Jiaxin
    Yonetani, Ryo
    Iqbal, Zahid
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7486 - 7492
  • [35] Mansour Y, 2020, Arxiv, DOI [arXiv:2002.10619, DOI 10.48550/ARXIV.2002.10619]
  • [36] McMahan HB, 2017, PR MACH LEARN RES, V54, P1273
  • [37] A Survey on Transfer Learning
    Pan, Sinno Jialin
    Yang, Qiang
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2010, 22 (10) : 1345 - 1359
  • [38] Pathak R., 2020, Advances in Neural Information Processing Systems, V33, P7057
  • [39] Poggio T., 2019, ONLINE LEARNING STAB
  • [40] Reddi S., 2020, Adaptive federated optimization