A Decentralized Proximal-Gradient Method With Network Independent Step-Sizes and Separated Convergence Rates

被引:148
作者
Li, Zhi [1 ]
Shi, Wei [2 ]
Yan, Ming [3 ]
机构
[1] Michigan State Univ, Dept Computat Math Sci & Engn, E Lansing, MI 48824 USA
[2] Princeton Univ, Dept Elect Engn, Princeton, NJ 08544 USA
[3] Michigan State Univ, Dept Math, Dept Computat Math Sci & Engn, E Lansing, MI 48824 USA
基金
美国国家科学基金会;
关键词
Decentralized optimization; proximal-gradient; convergence rates; network independent; DISTRIBUTED OPTIMIZATION; AVERAGE CONSENSUS; CONVEX-OPTIMIZATION; SUBGRADIENT METHODS; ALGORITHM;
D O I
10.1109/TSP.2019.2926022
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper proposes a novel proximal-gradient algorithm for a decentralized optimization problem with a composite objective containing smooth and nonsmooth terms. Specifically, the smooth and nonsmooth terms are dealt with by gradient and proximal updates, respectively. The proposed algorithm is closely related to a previous algorithm, PG-EXTRA (W. Shi, Q. Ling, G. Wu, and W. Yin, "A proximal gradient algorithm for decentralized composite optimization," IEEE Trans. Signal Process., vol. 63, no. 22, pp. 6013-6023,2015), but has a few advantages. First of all, agents use uncoordinated step-sizes, and the stable upper bounds on step-sizes are independent of network topologies. The step-sizes depend on local objective functions, and they can be as large as those of the gradient descent. Second, for the special case without nonsmooth terms, linear convergence can be achieved under the strong convexity assumption. The dependence of the convergence rate on the objective functions and the network are separated, and the convergence rate of the new algorithm is as good as one of the two convergence rates that match the typical rates for the general gradient descent and the consensus averaging. We provide numerical experiments to demonstrate the efficacy of the introduced algorithm and validate our theoretical discoveries.
引用
收藏
页码:4494 / 4506
页数:13
相关论文
共 58 条
[1]  
Alghunaim S. A., 2019, ARXIV190401196
[2]  
[Anonymous], FOUND TRENDS MACH LE
[3]  
[Anonymous], 2017, ARXIV171106785
[4]  
[Anonymous], 2013, INTRO LECT CONVEX OP, DOI DOI 10.1007/978-1-4419-8853-9
[5]  
[Anonymous], THESIS
[6]  
[Anonymous], 2017, ARXIV171200232
[7]   Distributed Spectrum Sensing for Cognitive Radio Networks by Exploiting Sparsity [J].
Bazerque, Juan Andres ;
Giannakis, Georgios B. .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2010, 58 (03) :1847-1862
[8]  
Bertsekas D., 2015, Parallel and Distributed Computation: Numerical Methods
[9]   Incremental proximal methods for large scale convex optimization [J].
Bertsekas, Dimitri P. .
MATHEMATICAL PROGRAMMING, 2011, 129 (02) :163-195
[10]   DISTRIBUTED ASYNCHRONOUS COMPUTATION OF FIXED-POINTS [J].
BERTSEKAS, DP .
MATHEMATICAL PROGRAMMING, 1983, 27 (01) :107-120