FedCust: Offloading hyperparameter customization for federated learning

被引:1
作者
Zawad, Syed [1 ]
Ma, Xiaolong [1 ]
Yi, Jun [1 ]
Li, Cheng [2 ]
Zhang, Minjia [2 ]
Yang, Lei [1 ]
Yan, Feng [3 ]
He, Yuxiong [2 ]
机构
[1] Univ Nevada, Dept Comp Sci & Engn, Reno, NV 89557 USA
[2] Microsoft, Bellevue, WA 98052 USA
[3] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
关键词
Federated learning; Hyperparameter optimization;
D O I
10.1016/j.peva.2024.102450
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose FedCust, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparametersto achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. FedCust uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, FedCust further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that FedCust achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.
引用
收藏
页数:19
相关论文
共 77 条
  • [51] Federated Learning With Cooperating Devices: A Consensus Approach for Massive IoT Networks
    Savazzi, Stefano
    Nicoli, Monica
    Rampa, Vittorio
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (05) : 4641 - 4654
  • [52] Smith V, 2018, Arxiv, DOI [arXiv:1705.10467, DOI 10.48550/ARXIV.1705.10467]
  • [53] Snoek J, 2015, PR MACH LEARN RES, V37, P2171
  • [54] Sun ZT, 2019, Arxiv, DOI arXiv:1911.07963
  • [55] Toward Personalized Federated Learning
    Tan, Alysa Ziying
    Yu, Han
    Cui, Lizhen
    Yang, Qiang
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 9587 - 9603
  • [56] Convergence properties of the expected improvement algorithm with fixed mean and covariance functions
    Vazquez, Emmanuel
    Bect, Julien
    [J]. JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2010, 140 (11) : 3088 - 3095
  • [57] Wang H, 2020, IEEE INFOCOM SER, P1698, DOI [10.1109/INFOCOM41043.2020.9155494, 10.1109/infocom41043.2020.9155494]
  • [58] Wang Jianyu, 2020, ADV NEURAL INFORM PR, V33
  • [59] Adaptive Federated Learning in Resource Constrained Edge Computing Systems
    Wang, Shiqiang
    Tuor, Tiffany
    Salonidis, Theodoros
    Leung, Kin K.
    Makaya, Christian
    He, Ting
    Chan, Kevin
    [J]. IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (06) : 1205 - 1221
  • [60] Poster: Exploiting Data Heterogeneity for Performance and Reliability in Federated Learning
    Wang, Yuanli
    Kumar, Dhruv
    Chandra, Abhishek
    [J]. 2020 IEEE/ACM SYMPOSIUM ON EDGE COMPUTING (SEC 2020), 2020, : 164 - 166