EXTRA: AN EXACT FIRST-ORDER ALGORITHM FOR DECENTRALIZED CONSENSUS OPTIMIZATION

被引:863
作者
Shi, Wei [1 ]
Ling, Qing [1 ]
Wu, Gang [1 ]
Yin, Wotao [2 ]
机构
[1] Univ Sci & Technol China, Dept Automat, Hefei 230026, Peoples R China
[2] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
基金
美国国家科学基金会;
关键词
consensus optimization; decentralized optimization; gradient method; linear convergence;
D O I
10.1137/14096668X
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Recently, there has been growing interest in solving consensus optimization problems in a multiagent network. In this paper, we develop a decentralized algorithm for the consensus optimization problem minimize(x is an element of Rp) (f) over bar (x) = (1)(n)Sigma(n)(i=1) f(i)(x), which is defined over a connected network of n agents, where each function f(i) is held privately by agent i and encodes the agent's data and objective. All the agents shall collaboratively find the minimizer while each agent can only communicate with its neighbors. Such a computation scheme avoids a data fusion center or long-distance communication and offers better load balance to the network. This paper proposes a novel decentralized exact first-order algorithm (abbreviated as EXTRA) to solve the consensus optimization problem. "Exact" means that it can converge to the exact solution. EXTRA uses a fixed, large step size, which can be determined independently of the network size or topology. The local variable of every agent i converges uniformly and consensually to an exact minimizer of (f) over bar. In contrast, the well-known decentralized gradient descent (DGD) method must use diminishing step sizes in order to converge to an exact minimizer. EXTRA and DGD have the same choice of mixing matrices and similar periteration complexity. EXTRA, however, uses the gradients of the last two iterates, unlike DGD which uses just that of the last iterate. EXTRA has the best known convergence rates among the existing synchronized first-order decentralized algorithms for minimizing convex Lipschitz-differentiable functions. Specifically, if the f(i)'s are convex and have Lipschitz continuous gradients, EXTRA has an ergodic convergence rate O(1/k) in terms of the first-order optimality residual. In addition, as long as (f) over bar is (restricted) strongly convex (not all individual f(i)'s need to be so), EXTRA converges to an optimal solution at a linear rate O(C-k) for some constant C > 1.
引用
收藏
页码:944 / 966
页数:23
相关论文
共 37 条
  • [1] [Anonymous], ARXIV14064834
  • [2] [Anonymous], 2013, ARXIV13078254
  • [3] [Anonymous], 2013, ARXIV13121085
  • [4] [Anonymous], 2013, ARXIV13107063
  • [5] Group-Lasso on Splines for Spectrum Cartography
    Bazerque, Juan Andres
    Mateos, Gonzalo
    Giannakis, Georgios B.
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2011, 59 (10) : 4648 - 4663
  • [6] Distributed Spectrum Sensing for Cognitive Radio Networks by Exploiting Sparsity
    Bazerque, Juan Andres
    Giannakis, Georgios B.
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2010, 58 (03) : 1847 - 1862
  • [7] Boyd S, 2004, SIAM REV, V46, P667, DOI [10.1137/S0036144503423264, 10.1137/s0036144503423264]
  • [8] Multi-Agent Distributed Optimization via Inexact Consensus ADMM
    Chang, Tsung-Hui
    Hong, Mingyi
    Wang, Xiangfeng
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2015, 63 (02) : 482 - 497
  • [9] Chen I.-A., 2012, Ph.D. thesis)
  • [10] Gossip Algorithms for Distributed Signal Processing
    Dimakis, Alexandros G.
    Kar, Soummya
    Moura, Jose M. F.
    Rabbat, Michael G.
    Scaglione, Anna
    [J]. PROCEEDINGS OF THE IEEE, 2010, 98 (11) : 1847 - 1864