FedCust: Offloading hyperparameter customization for federated learning

被引:1
作者
Zawad, Syed [1 ]
Ma, Xiaolong [1 ]
Yi, Jun [1 ]
Li, Cheng [2 ]
Zhang, Minjia [2 ]
Yang, Lei [1 ]
Yan, Feng [3 ]
He, Yuxiong [2 ]
机构
[1] Univ Nevada, Dept Comp Sci & Engn, Reno, NV 89557 USA
[2] Microsoft, Bellevue, WA 98052 USA
[3] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
关键词
Federated learning; Hyperparameter optimization;
D O I
10.1016/j.peva.2024.102450
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose FedCust, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparametersto achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. FedCust uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, FedCust further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that FedCust achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.
引用
收藏
页数:19
相关论文
共 77 条
  • [1] Agarwal N, 2021, ADV NEUR IN
  • [2] Genetic CFL: Hyperparameter Optimization in Clustered Federated Learning
    Agrawal, Shaashwat
    Sarkar, Sagnik
    Alazab, Mamoun
    Maddikunta, Praveen Kumar Reddy
    Gadekallu, Thippa Reddy
    Quoc-Viet Pham
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2021, 2021
  • [3] Aledhari M, 2020, IEEE ACCESS, V8, P140699, DOI [10.1109/ACCESS.2020.3013541, 10.1109/access.2020.3013541]
  • [4] [Anonymous], 2022, The android profiler
  • [5] Balcan M., 2020, INT WORKSH FED LEARN
  • [6] Bergstra J, 2012, J MACH LEARN RES, V13, P281
  • [7] Bonawitz K., 2019, ARXIV190201046, DOI 10.48550/arXiv.1902.01046
  • [8] Practical Secure Aggregation for Privacy-Preserving Machine Learning
    Bonawitz, Keith
    Ivanov, Vladimir
    Kreuter, Ben
    Marcedone, Antonio
    McMahan, H. Brendan
    Patel, Sarvar
    Ramage, Daniel
    Segal, Aaron
    Seth, Karn
    [J]. CCS'17: PROCEEDINGS OF THE 2017 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2017, : 1175 - 1191
  • [9] Federated learning with hierarchical clustering of local updates to improve training on non-IID data
    Briggs, Christopher
    Fan, Zhong
    Andras, Peter
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [10] Caldas S, 2018, Leaf: A benchmark for federated settings, DOI DOI 10.48550/ARXIV.1812.01097