GraphCS: Graph-based client selection for heterogeneity in federated learning

被引:8
作者
Chang, Tao [1 ]
Li, Li [2 ]
Wu, MeiHan [1 ]
Yu, Wei [3 ]
Wang, Xiaodong [1 ]
Xu, ChengZhong [2 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Key Lab Parallel & Distributed Comp, Changsha, Peoples R China
[2] Univ Macau, State Key Lab Internet Things Smart City, Taipa, Peoples R China
[3] China Elect Technol Grp Corp, Res Inst 30, Chengdu, Peoples R China
关键词
Federated learning; Client selection; Heterogeneity; ALGORITHMS;
D O I
10.1016/j.jpdc.2023.03.003
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Federated Learning coordinates many mobile devices to train an artificial intelligence model while preserving data privacy collaboratively. Mobile devices are usually equipped with totally different hardware configurations, leading to various training capabilities. At the same time, the distribution of the local training data is highly heterogeneous across different clients. Randomly selecting the clients to participate in the training process results in poor model performance and low system efficiency. In this paper, we propose GraphCS, a graph-based client selection framework for heterogeneity in Federated Learning. GraphCS first measures the distribution coupling across the clients via the model gradients. After that, it divides the clients into different groups according to the diversity of the local datasets. At the same time, it well estimates the runtime training capability of each client by jointly considering the hardware configuration and resource contention caused by the concurrently running apps. With the distribution coupling information and runtime training capability, GraphCS selects the best clients in order to well balance the model accuracy and overall training progress. We evaluate the performance of GraphCS with mobile devices with different hardware configurations on various datasets. The experiment results show that our approach improves model accuracy up to 45.69%. Meanwhile, it reduces communication and computation overhead 87.35% and 89.48% at best, respectively. Furthermore, GraphCS accelerates the overall training process up to 35x. (c) 2023 Elsevier Inc. All rights reserved.
引用
收藏
页码:131 / 143
页数:13
相关论文
共 40 条
[1]  
Balakrishnan R., 2021, DIVERSE CLIENT SELEC
[2]  
Bonawitz K., 2019, ARXIV
[3]   Practical Secure Aggregation for Privacy-Preserving Machine Learning [J].
Bonawitz, Keith ;
Ivanov, Vladimir ;
Kreuter, Ben ;
Marcedone, Antonio ;
McMahan, H. Brendan ;
Patel, Sarvar ;
Ramage, Daniel ;
Segal, Aaron ;
Seth, Karn .
CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, :1175-1191
[4]  
Canziani Alfredo, 2016, arXiv
[5]  
Cho K., 2014, C EMP METH NAT LANG, DOI DOI 10.3115/V1/W14-4012
[6]  
Collins J., 2012, CLEF 2012 EV LABS WO
[7]  
Deng YB, 2019, Arxiv, DOI [arXiv:1904.09274, 10.13140/RG.2.2.15012.12167]
[8]   The Algorithmic Foundations of Differential Privacy [J].
Dwork, Cynthia ;
Roth, Aaron .
FOUNDATIONS AND TRENDS IN THEORETICAL COMPUTER SCIENCE, 2013, 9 (3-4) :211-406
[9]  
E.D.D. Team, 2016, DL4J DEEP LEARN JAV
[10]   Discovering faster matrix multiplication algorithms with reinforcement learning [J].
Fawzi, Alhussein ;
Balog, Matej ;
Huang, Aja ;
Hubert, Thomas ;
Romera-Paredes, Bernardino ;
Barekatain, Mohammadamin ;
Novikov, Alexander ;
Ruiz, Francisco J. R. ;
Schrittwieser, Julian ;
Swirszcz, Grzegorz ;
Silver, David ;
Hassabis, Demis ;
Kohli, Pushmeet .
NATURE, 2022, 610 (7930) :47-+