FedCust: Offloading hyperparameter customization for federated learning

被引:1
作者
Zawad, Syed [1 ]
Ma, Xiaolong [1 ]
Yi, Jun [1 ]
Li, Cheng [2 ]
Zhang, Minjia [2 ]
Yang, Lei [1 ]
Yan, Feng [3 ]
He, Yuxiong [2 ]
机构
[1] Univ Nevada, Dept Comp Sci & Engn, Reno, NV 89557 USA
[2] Microsoft, Bellevue, WA 98052 USA
[3] Univ Houston, Dept Comp Sci, Houston, TX 77204 USA
关键词
Federated learning; Hyperparameter optimization;
D O I
10.1016/j.peva.2024.102450
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) is a new machine learning paradigm that enables training models collaboratively across clients without sharing private data. In FL, data is non-uniformly distributed among clients (i.e., data heterogeneity) and cannot be redistributed nor monitored like in conventional machine learning due to privacy constraints. Such data heterogeneity and privacy requirements bring new challenges for learning hyperparameter optimization as the training dynamics change across clients even within the same training round and they are difficult to be measured due to privacy. The state-of-the-art in hyperparameter customization can greatly improve FL model accuracy but also incur significant computing overheads and power consumption on client devices, and slowdown the training process. To address the prohibitively expensive cost challenge, we explore the possibility of offloading hyperparameter customization to servers. We propose FedCust, a framework that offloads expensive hyperparameter customization cost from the client devices to the central server without violating privacy constraints. Our key discovery is that it is not necessary to do hyperparameter customization for every client, and clients with similar data heterogeneity can use the same hyperparametersto achieve good training performance. We propose heterogeneity measurement metrics for clustering clients into groups such that clients within the same group share hyperparameters. FedCust uses the proxy data from initial model design to emulate different heterogeneity groups and perform hyperparameter customization on the server side without accessing client data nor information. To make the hyperparameter customization scalable, FedCust further employs a Bayesian-strengthened tuner to significantly accelerates the hyperparameter customization speed. Extensive evaluation demonstrates that FedCust achieves up to 7/2/4/4/6% better accuracy than the widely adopted one-size-fits-all approach on popular FL benchmarks FEMNIST, Shakespeare, Cifar100, Cifar10, and Fashion-MNIST respectively, while being scalable and reducing computation, memory, and energy consumption on the client devices, without compromising privacy constraints.
引用
收藏
页数:19
相关论文
共 77 条
  • [11] The Art of Designing Remote IoT Devices-Technologies and Strategies for a Long Battery Life
    Callebaut, Gilles
    Leenders, Guus
    Van Mulders, Jarne
    Ottoy, Geoffrey
    De Strycker, Lieven
    van der Perre, Liesbet
    [J]. SENSORS, 2021, 21 (03) : 1 - 37
  • [12] TiFL: A Tier-based Federated Learning System
    Chai, Zheng
    Ali, Ahsan
    Zawad, Syed
    Treux, Stacey
    Anwar, Ali
    Barcaldo, Nathalie
    Zhou, Yi
    Ludwig, Heiko
    Yan, Feng
    Cheng, Yue
    [J]. PROCEEDINGS OF THE 29TH INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2020, 2020, : 125 - 136
  • [13] Chen TY, 2020, Arxiv, DOI arXiv:2007.06081
  • [14] Dai Z, 2021, ADV NEUR IN, V34
  • [15] Dai ZW, 2020, ADV NEUR IN, V33
  • [16] Ding Guanyu, 2022, Cloud Computing - CLOUD 2022: 15th International Conference, Held as Part of the Services Conference Federation, SCF 2022, Proceedings. Lecture Notes in Computer Science (13731), P13, DOI 10.1007/978-3-031-23498-9_2
  • [17] Fallah A, 2020, Arxiv, DOI arXiv:2002.07948
  • [18] Ghosh A, 2019, Arxiv, DOI [arXiv:1906.06629, DOI 10.48550/ARXIV.1906.06629]
  • [19] Guha N, 2019, Arxiv, DOI arXiv:1902.11175
  • [20] Efficient and flexible management for industrial Internet of Things: A federated learning approach
    Guo, Yinghao
    Zhao, Zichao
    He, Ke
    Lai, Shiwei
    Xia, Junjuan
    Fan, Lisheng
    [J]. COMPUTER NETWORKS, 2021, 192