FedCust: Offloading hyperparameter customization for federated learning

被引:2
作者
Zawad, Syed [1 ]
Ma, Xiaolong [1 ]
Yi, Jun [1 ]
Li, Cheng [2 ]
Zhang, Minjia [2 ]
Yang, Lei [1 ]
Yan, Feng [3 ]
He, Yuxiong [2 ]
机构
[1] Univ Nevada, Dept Comp Sci & Engn, Reno, NV 89557 USA
[2] Microsoft, Bellevue, WA 98052 USA
[3] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
关键词
Federated learning; Hyperparameter optimization;
D O I
10.1016/j.peva.2024.102450
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose FedCust, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparametersto achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. FedCust uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, FedCust further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that FedCust achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.
引用
收藏
页数:19
相关论文
共 77 条
[11]   TiFL: A Tier-based Federated Learning System [J].
Chai, Zheng ;
Ali, Ahsan ;
Zawad, Syed ;
Treux, Stacey ;
Anwar, Ali ;
Barcaldo, Nathalie ;
Zhou, Yi ;
Ludwig, Heiko ;
Yan, Feng ;
Cheng, Yue .
PROCEEDINGS OF THE 29TH INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2020, 2020, :125-136
[12]  
Chen TY, 2020, Arxiv, DOI arXiv:2007.06081
[13]  
Dai Z., 2020, Adv. Neural Inf. Process. Syst., V33
[14]  
Dai Z, 2021, ADV NEUR IN, V34
[15]  
Ding Guanyu, 2022, Cloud Computing - CLOUD 2022: 15th International Conference, Held as Part of the Services Conference Federation, SCF 2022, Proceedings. Lecture Notes in Computer Science (13731), P13, DOI 10.1007/978-3-031-23498-9_2
[16]  
Fallah A, 2020, Arxiv, DOI arXiv:2002.07948
[17]  
Ghosh A, 2019, Arxiv, DOI arXiv:1906.06629
[18]  
Guha N, 2019, Arxiv, DOI arXiv:1902.11175
[19]   Efficient and flexible management for industrial Internet of Things: A federated learning approach [J].
Guo, Yinghao ;
Zhao, Zichao ;
He, Ke ;
Lai, Shiwei ;
Xia, Junjuan ;
Fan, Lisheng .
COMPUTER NETWORKS, 2021, 192
[20]  
Hard A, 2019, Arxiv, DOI arXiv:1811.03604