Distributed Solution of Large-Scale Linear Systems via Accelerated Projection-Based Consensus

被引:21
作者
Azizan-Ruhi, Navid [1 ]
Lahouti, Farshad [2 ]
Avestimehr, Amir Salman [3 ]
Hassibi, Babak [2 ]
机构
[1] CALTECH, Dept Comp & Math Sci, Pasadena, CA 91125 USA
[2] CALTECH, Dept Elect Engn, Pasadena, CA 91125 USA
[3] Univ Southern Calif, Dept Elect Engn, Los Angeles, CA 90007 USA
基金
美国国家科学基金会;
关键词
System of linear equations; distributed computing; big data; consensus; optimization; CONVERGENCE; ALGORITHM; ADMM;
D O I
10.1109/TSP.2019.2917855
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Solving a large-scale system of linear equations is a key step at the heart of many algorithms in scientific computing, machine learning, and beyond. When the problem dimension is large, computational and/or memory constraints make it desirable, or even necessary, to perform the task in a distributed fashion. In this paper, we consider a common scenario in which a taskmaster intends to solve a large-scale system of linear equations by distributing subsets of the equations among a number of computing machines/cores. We propose a new algorithm called Accelerated Projection-based Consensus, in which at each iteration every machine updates its solution by adding a scaled version of the projection of an error signal onto the nullspace of its system of equations, and the taskmaster conducts an averaging over the solutions with momentum. The convergence behavior of the proposed algorithm is analyzed in detail and analytically shown to compare favorably with the convergence rate of alternative distributed methods, namely distributed gradient descent, distributed versions of Nesterov's accelerated gradient descent and heavy-ball method, the block Cimmino method, and Alternating Direction Method of Multipliers. On randomly chosen linear systems, as well as on real-world data sets, the proposed method offers significant speed-up relative to all the aforementioned methods. Finally, our analysis suggests a novel variation of the distributed heavy-ball method, which employs a particular distributed preconditioning and achieves the same theoretical convergence rate as that in the proposed consensus-based method.
引用
收藏
页码:3806 / 3817
页数:12
相关论文
共 25 条
[1]  
[Anonymous], FOUND TRENDS MACH LE
[2]  
[Anonymous], 2007, MATRIX MARKET
[3]  
[Anonymous], 2014, P INT C MACH LEARN
[4]  
[Anonymous], 1964, COMP MATH MATH PHYS+
[5]   A BLOCK PROJECTION METHOD FOR SPARSE MATRICES [J].
ARIOLI, M ;
DUFF, I ;
NOAILLES, J ;
RUIZ, D .
SIAM JOURNAL ON SCIENTIFIC AND STATISTICAL COMPUTING, 1992, 13 (01) :47-70
[6]  
Azizan-Ruhi N, 2018, 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), P6358, DOI 10.1109/ICASSP.2018.8462630
[7]   ROW PROJECTION METHODS FOR LARGE NONSYMMETRIC LINEAR-SYSTEMS [J].
BRAMLEY, R ;
SAMEH, A .
SIAM JOURNAL ON SCIENTIFIC AND STATISTICAL COMPUTING, 1992, 13 (01) :168-193
[8]   On the Global and Linear Convergence of the Generalized Alternating Direction Method of Multipliers [J].
Deng, Wei ;
Yin, Wotao .
JOURNAL OF SCIENTIFIC COMPUTING, 2016, 66 (03) :889-916
[9]   THE AUGMENTED BLOCK CIMMINO DISTRIBUTED METHOD [J].
Duff, Iain S. ;
Guivarch, Ronan ;
Ruiz, Daniel ;
Zenadi, Mohamed .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2015, 37 (03) :A1248-A1269
[10]   ON THE O(1/n) CONVERGENCE RATE OF THE DOUGLAS-RACHFORD ALTERNATING DIRECTION METHOD [J].
He, Bingsheng ;
Yuan, Xiaoming .
SIAM JOURNAL ON NUMERICAL ANALYSIS, 2012, 50 (02) :700-709