Achieving Linear Speedup in Asynchronous Federated Learning With Heterogeneous Clients

被引:0
作者
Wang, Xiaolu [1 ]
Li, Zijian [2 ]
Jin, Shi [3 ]
Zhang, Jun [2 ]
机构
[1] East China Normal Univ, Software Engn Inst, Shanghai 200050, Peoples R China
[2] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Hong Kong, Peoples R China
[3] Southeast Univ, Natl Mobile Commun Res Lab, Nanjing 211102, Peoples R China
关键词
Training; Servers; Convergence; Computational modeling; Scalability; Federated learning; Stochastic processes; Asynchronous federated learning; distributed optimization; edge machine learning; linear speedup; system heterogeneity; CONVERGENCE ANALYSIS;
D O I
10.1109/TMC.2024.3461852
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) is an emerging distributed training paradigm that aims to learn a common global model without exchanging or transferring the data that are stored locally at different clients. The Federated Averaging (FedAvg)-based algorithms have gained substantial popularity in FL to reduce the communication overhead, where each client conducts multiple localized iterations before communicating with a central server. In this paper, we focus on FL where the clients have diverse computation and/or communication capabilities. Under this circumstance, FedAvg can be less efficient since it requires all clients that participate in the global aggregation in a round to initiate iterations from the latest global model, and thus the synchronization among fast clients and straggler clients can severely slow down the overall training process. To address this issue, we propose an efficient asynchronous federated learning (AFL) framework called Delayed Federated Averaging (DeFedAvg). In DeFedAvg, the clients are allowed to perform local training with different stale global models at their own paces. Theoretical analyses demonstrate that DeFedAvg achieves asymptotic convergence rates that are on par with the results of FedAvg for solving nonconvex problems. More importantly, DeFedAvg is the first AFL algorithm that provably achieves the desirable linear speedup property, which indicates its high scalability. Additionally, we carry out extensive numerical experiments using real datasets to validate the efficiency and scalability of our approach when training deep neural networks.
引用
收藏
页码:435 / 448
页数:14
相关论文
共 58 条
  • [1] Federated Learning in Edge Computing: A Systematic Survey
    Abreha, Haftay Gebreslasie
    Hayajneh, Mohammad
    Serhani, Mohamed Adel
    [J]. SENSORS, 2022, 22 (02)
  • [2] Agarwal A, 2012, IEEE DECIS CONTR P, P5451, DOI 10.1109/CDC.2012.6426626
  • [3] Arjevani Y, 2020, PR MACH LEARN RES, V117, P111
  • [4] Advances in Asynchronous Parallel and Distributed Optimization
    Assran, By Mahmoud
    Aytekin, Arda
    Feyzmahdavian, Hamid Reza
    Johansson, Mikael
    Rabbat, Michael G.
    [J]. PROCEEDINGS OF THE IEEE, 2020, 108 (11) : 2013 - 2031
  • [5] Avdiukhin Dmitrii, 2021, P MACHINE LEARNING R, P425
  • [6] Bertsekas D. P., 1997, Introduction to Probability
  • [7] Asynchronous Online Federated Learning for Edge Devices with Non-IID Data
    Chen, Yujing
    Ning, Yue
    Slawski, Martin
    Rangwala, Huzefa
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 15 - 24
  • [8] Cho Y.J., 2020, ARXIV
  • [9] Dutta S, 2018, PR MACH LEARN RES, V84
  • [10] Fraboni Y, 2023, J MACH LEARN RES, V24