GraphCS: Graph-based client selection for heterogeneity in federated learning

被引:8
作者
Chang, Tao [1 ]
Li, Li [2 ]
Wu, MeiHan [1 ]
Yu, Wei [3 ]
Wang, Xiaodong [1 ]
Xu, ChengZhong [2 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Key Lab Parallel & Distributed Comp, Changsha, Peoples R China
[2] Univ Macau, State Key Lab Internet Things Smart City, Taipa, Peoples R China
[3] China Elect Technol Grp Corp, Res Inst 30, Chengdu, Peoples R China
关键词
Federated learning; Client selection; Heterogeneity; ALGORITHMS;
D O I
10.1016/j.jpdc.2023.03.003
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Federated Learning coordinates many mobile devices to train an artificial intelligence model while preserving data privacy collaboratively. Mobile devices are usually equipped with totally different hardware configurations, leading to various training capabilities. At the same time, the distribution of the local training data is highly heterogeneous across different clients. Randomly selecting the clients to participate in the training process results in poor model performance and low system efficiency. In this paper, we propose GraphCS, a graph-based client selection framework for heterogeneity in Federated Learning. GraphCS first measures the distribution coupling across the clients via the model gradients. After that, it divides the clients into different groups according to the diversity of the local datasets. At the same time, it well estimates the runtime training capability of each client by jointly considering the hardware configuration and resource contention caused by the concurrently running apps. With the distribution coupling information and runtime training capability, GraphCS selects the best clients in order to well balance the model accuracy and overall training progress. We evaluate the performance of GraphCS with mobile devices with different hardware configurations on various datasets. The experiment results show that our approach improves model accuracy up to 45.69%. Meanwhile, it reduces communication and computation overhead 87.35% and 89.48% at best, respectively. Furthermore, GraphCS accelerates the overall training process up to 35x. (c) 2023 Elsevier Inc. All rights reserved.
引用
收藏
页码:131 / 143
页数:13
相关论文
共 40 条
[21]  
Li L, 2020, UK EU CHINA MILLIMET, DOI [10.1109/ucmmt49983.2020.9295999, 10.1109/ICPICS50287.2020.9202096]
[22]  
Li T., 2020, PROC MACH LEARN SYST, V2, P429
[23]  
Li W., MACHINE LEARNING MED
[24]   Gaussian elimination is not optimal, revisited [J].
Macedo, Hugo Daniel .
JOURNAL OF LOGICAL AND ALGEBRAIC METHODS IN PROGRAMMING, 2016, 85 (05) :999-1010
[25]  
McMahan HB, 2017, PR MACH LEARN RES, V54, P1273
[26]  
Pedia W., PERF LINUX PROFILING
[27]   Solution methods and computational investigations for the linear bottleneck assignment problem [J].
Pferschy, U .
COMPUTING, 1997, 59 (03) :237-258
[28]  
Shepard Clayton, 2010, Performance Evaluation Review, V38, P15, DOI 10.1145/1925019.1925024
[29]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[30]  
Singh R. P., 2011, 2011 IEEE 3rd International Conference on Communication Software and Networks (ICCSN 2011), P591, DOI 10.1109/ICCSN.2011.6013975