Can hierarchical client clustering mitigate the data heterogeneity effect in federated learning?

被引:1
作者
Lee, Seungjun [1 ]
Yu, Miri [1 ]
Yoon, Daegun [1 ]
Oh, Sangyoon [1 ]
机构
[1] Ajou Univ, Dept Artificial Intelligence, Suwon, South Korea
来源
2023 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW | 2023年
基金
新加坡国家研究基金会;
关键词
federated learning; data heterogeneity; hierarchical aggregation; client clustering;
D O I
10.1109/IPDPSW59300.2023.00134
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) was proposed for training a deep neural network model using millions of user data. The technique has attracted considerable attention owing to its privacy-preserving characteristic. However, two major challenges exist. The first is the limitation of simultaneously participating clients. If the number of clients increases, the single parameter server easily becomes a bottleneck and is prone to have stragglers. The second is data heterogeneity, which adversely affects the accuracy of the global model. Because data should remain at user devices to preserve privacy, we cannot use data shuffling, which is used to homogenize training data in traditional distributed deep learning. We propose a client clustering and model aggregation method, CCFed, to increase the number of simultaneously participating clients and mitigate the data heterogeneity problem. CCFed improves the learning performance using set partition modeling to let data be evenly distributed between clusters and mitigate the effect of a non-IID environment. Experiments show that we can achieve a 2.7-14% higher accuracy using CCFed compared with FedAvg, where CCFed requires approximately 50% less number of rounds compared with FedAvg training on benchmark datasets.
引用
收藏
页码:799 / 808
页数:10
相关论文
共 31 条
[1]  
Bhuyan Neelkamal, 2022, 2022 14th International Conference on COMmunication Systems & NETworkS (COMSNETS), P779, DOI 10.1109/COMSNETS53615.2022.9668435
[2]   Towards Federated Learning using FaaS Fabric [J].
Chadha, Mohak ;
Jindal, Anshul ;
Gerndt, Michael .
PROCEEDINGS OF THE 2020 SIXTH INTERNATIONAL WORKSHOP ON SERVERLESS COMPUTING (WOSC '20), 2020, :49-54
[3]   Asynchronous Federated Learning for Sensor Data with Concept Drift [J].
Chen, Yujing ;
Chai, Zheng ;
Cheng, Yue ;
Rangwala, Huzefa .
2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, :4822-4831
[4]   Fundamental Technologies in Modern Speech Recognition [J].
Furui, Sadaoki ;
Deng, Li ;
Gales, Mark ;
Ney, Hermann ;
Tokuda, Keiichi .
IEEE SIGNAL PROCESSING MAGAZINE, 2012, 29 (06) :16-17
[5]  
Grafberger A, 2021, Arxiv, DOI arXiv:2111.03396
[6]   BOUNDS ON MULTIPROCESSING TIMING ANOMALIES [J].
GRAHAM, RL .
SIAM JOURNAL ON APPLIED MATHEMATICS, 1969, 17 (02) :416-&
[7]   A Cluster-Driven Adaptive Training Approach for Federated Learning [J].
Jeong, Younghwan ;
Kim, Taeyoon .
SENSORS, 2022, 22 (18)
[8]  
Karimireddy SP, 2020, PR MACH LEARN RES, V119
[9]   Federated Learning for Edge Networks: Resource Optimization and Incentive Mechanism [J].
Khan, Latif U. ;
Pandey, Shashi Raj ;
Tran, Nguyen H. ;
Saad, Walid ;
Han, Zhu ;
Nguyen, Minh N. H. ;
Hong, Choong Seon .
IEEE COMMUNICATIONS MAGAZINE, 2020, 58 (10) :88-93
[10]  
Korf R., 2010, P SOCS, V1, P71, DOI DOI 10.1609/SOCS.V1I1.18172