Decentralized Over-the-Air Federated Learning by Second-Order Optimization Method

被引:4
作者
Yang, Peng [1 ,2 ]
Jiang, Yuning [3 ]
Wen, Dingzhu [4 ]
Wang, Ting [1 ,2 ]
Jones, Colin N. [3 ]
Shi, Yuanming [4 ]
机构
[1] East China Normal Univ, Shanghai Key Lab Trustworthy Comp, Shanghai 200062, Peoples R China
[2] East China Normal Univ, MoE Engn Res Ctr Soft ware Hardware Codesign Techn, Shanghai 200062, Peoples R China
[3] Ecole Polytech Fed Lausanne, Automat Control Lab, CH-1015 Lausanne, Switzerland
[4] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai 201210, Peoples R China
关键词
Decentralized federated learning; over-the-air computation; second-order optimization method; COMMUNICATION; COMPUTATION; CHALLENGES; ALGORITHMS; NETWORKS; PRIVACY;
D O I
10.1109/TWC.2023.3327610
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Federated learning (FL) is an emerging technique that enables privacy-preserving distributed learning. Most related works focus on centralized FL, which leverages the coordination of a parameter server to implement local model aggregation. However, this scheme heavily relies on the parameter server, which could cause scalability, communication, and reliability issues. To tackle these problems, decentralized FL, where information is shared through gossip, starts to attract attention. Nevertheless, current research mainly relies on first-order optimization methods that have a relatively slow convergence rate, which leads to excessive communication rounds in wireless networks. To design communication-efficient decentralized FL, we propose a novel over-the-air decentralized second-order federated algorithm. Benefiting from the fast convergence rate of the second-order method, total communication rounds are significantly reduced. Meanwhile, owing to the low-latency model aggregation enabled by over-the-air computation, the communication overheads in each round can also be greatly decreased. The convergence behavior of our approach is then analyzed. The result reveals an error term, which involves a cumulative noise effect, in each iteration. To mitigate the impact of this error term, we conduct system optimization from the perspective of the accumulative term and the individual term, respectively. Numerical experiments demonstrate the superiority of our proposed approach and the effectiveness of system optimization.
引用
收藏
页码:5632 / 5647
页数:16
相关论文
共 56 条
  • [1] Alghunaim S. A., 2019, Advances in Neural Information Processing Systems, P2848
  • [2] NEWTON-LIKE METHOD WITH DIAGONAL CORRECTION FOR DISTRIBUTED OPTIMIZATION
    Bajovic, Dragana
    Jakovetic, Dusan
    Krejic, Natasa
    Jerinkic, Natasa Krklec
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 1171 - 1203
  • [3] Decentralized federated learning for extended sensing in 6G connected vehicles
    Barbieri, Luca
    Savazzi, Stefano
    Brambilla, Mattia
    Nicoli, Monica
    [J]. VEHICULAR COMMUNICATIONS, 2022, 33
  • [4] Communication Efficient Decentralized Learning Over Bipartite Graphs
    Ben Issaid, Chaouki
    Elgabli, Anis
    Park, Jihong
    Bennis, Mehdi
    Debbah, Merouane
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (06) : 4150 - 4167
  • [5] A Uniform-Forcing Transceiver Design for Over-the-Air Function Computation
    Chen, Li
    Qin, Xiaowei
    Wei, Guo
    [J]. IEEE WIRELESS COMMUNICATIONS LETTERS, 2018, 7 (06) : 942 - 945
  • [6] A Joint Learning and Communications Framework for Federated Learning Over Wireless Networks
    Chen, Mingzhe
    Yang, Zhaohui
    Saad, Walid
    Yin, Changchuan
    Poor, H. Vincent
    Cui, Shuguang
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (01) : 269 - 283
  • [7] Chen Yuzhou, 2021, P MACHINE LEARNING R, V139
  • [8] Daneshmand A., 2021, P MACHINE LEARNING R, P2398
  • [9] Decentralized Quasi-Newton Methods
    Eisen, Mark
    Mokhtari, Aryan
    Ribeiro, Alejandro
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2017, 65 (10) : 2613 - 2628
  • [10] Eisen M, 2016, IEEE DECIS CONTR P, P1951, DOI 10.1109/CDC.2016.7798550