Consensus and Diffusion for First-Order Distributed Optimization Over Multi-Hop Network

被引:0
作者
Wu, Mou [1 ,2 ]
Liao, Haibin [3 ]
Tan, Liansheng [4 ]
Jin, Guonian [1 ]
Zhong, Liangji [1 ]
Xiao, Yonggang [1 ]
Wang, Zhiyong [1 ]
机构
[1] Hubei Univ Sci & Technol, Sch Comp Sci & Technol, Xianning 437100, Peoples R China
[2] Hubei Univ Sci & Technol, Lab Optoelect Informat & Intelligent Control, Xianning 437100, Peoples R China
[3] Wuhan Text Univ, Sch Elect & Elect Engn, Wuhan 430200, Peoples R China
[4] Cent China Normal Univ, Dept Comp Sci, Wuhan 430079, Peoples R China
基金
中国国家自然科学基金;
关键词
Distributed optimization; gradient descent; consensus; diffusion; multi-hop network; GRADIENT METHODS; CONVERGENCE; COMMUNICATION; ALGORITHMS;
D O I
10.1109/ACCESS.2023.3297112
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distributed optimization is a powerful paradigm to solve various problems in machine learning over networked systems. Existing first-order optimization methods perform cheap gradient descent by exchanging information per iteration only with single-hop neighbours in a network. However, in many agent networks such as sensor and robotic networks, it is prevalent that each agent can interact with other agents over multi-hop communication. Therefore, distributed optimization method over multi-hop networks is an important but overlooked topic that clearly needs to be developed. Motivated by this observation, in this paper, we apply multi-hop transmission to the outstanding distributed gradient descent (DGD) method and propose two typical versions (i.e., consensus and diffusion) of multi-hop DGD method, which we call CM-DGD and DM-DGD, respectively. Theoretically, we present the convergence guarantee of the proposed methods under mild assumptions. Moreover, we confirm that multi-hop strategy results in higher probability to improve the spectral gap of the underlying network, which has been shown to be a critical factor improving performance of distributed optimizations, thus achieves better convergence metrics. Experimentally, two distributed machine learning problems are picked to verify the theoretical analysis and show the effectiveness of CM-DGD and DM-DGD by using synthesized and real data sets.
引用
收藏
页码:76913 / 76925
页数:13
相关论文
共 52 条
  • [1] NEWTON-LIKE METHOD WITH DIAGONAL CORRECTION FOR DISTRIBUTED OPTIMIZATION
    Bajovic, Dragana
    Jakovetic, Dusan
    Krejic, Natasa
    Jerinkic, Natasa Krklec
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 1171 - 1203
  • [2] On the Convergence of Nested Decentralized Gradient Methods With Multiple Consensus and Gradient Steps
    Berahas, Albert S.
    Bollapragada, Raghu
    Wei, Ermin
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 (69) : 4192 - 4203
  • [3] Berahas AS, 2019, IEEE DECIS CONTR P, P1519, DOI 10.1109/CDC40024.2019.9029541
  • [4] Balancing Communication and Computation in Distributed Optimization
    Berahas, Albert S.
    Bollapragada, Raghu
    Keskar, Nitish Shirish
    Wei, Ermin
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2019, 64 (08) : 3141 - 3155
  • [5] Bertsekas D. P., 2009, Convex optimization theory
  • [6] A convergent incremental gradient method with a constant step size
    Blatt, Doron
    Hero, Alfred O.
    Gauchman, Hillel
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2007, 18 (01) : 29 - 51
  • [7] Boyd S, 2005, IEEE INFOCOM SER, P1653
  • [8] Fastest mixing Markov chain on a path
    Boyd, S
    Diaconis, P
    Sun, J
    Xiao, L
    [J]. AMERICAN MATHEMATICAL MONTHLY, 2006, 113 (01) : 70 - 74
  • [9] Multihop diversity in wireless relaying channels
    Boyer, J
    Falconer, DD
    Yanikomeroglu, H
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2004, 52 (10) : 1820 - 1830
  • [10] Distributed Gradient Descent Algorithm Robust to an Arbitrary Number of Byzantine Attackers
    Cao, Xinyang
    Lai, Lifeng
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (22) : 5850 - 5864