Randomized Block Proximal Methods for Distributed Stochastic Big-Data Optimization

被引:4
作者
Farina, Francesco [1 ]
Notarstefano, Giuseppe [2 ]
机构
[1] GlaxoSmithKline, Artificial Intelligence & Machine Learning Grp, London TW8 9GS, England
[2] Univ Bologna, Dept Elect Elect & Informat Engn G Marconi, I-40136 Bologna, Italy
基金
欧洲研究理事会;
关键词
Optimization; Convergence; Linear programming; Stochastic processes; Distributed algorithms; Approximation algorithms; Heuristic algorithms; Big data applications; distributed algorithms; optimization methods; Stochastic systems; CONVERGENCE ANALYSIS; DESCENT METHODS; ALGORITHMS; NONSMOOTH; CONSENSUS;
D O I
10.1109/TAC.2020.3027647
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this article, we introduce a class of novel distributed algorithms for solving stochastic big-data convex optimization problems over directed graphs. In the addressed set-up, the dimension of the decision variable can be extremely high and the objective function can be nonsmooth. The general algorithm consists of two main steps: a consensus step and an update on a single block of the optimization variable, which is then broadcast to neighbors. Three special instances of the proposed method, involving particular problem structures, are then presented. In the general case, the convergence of a dynamic consensus algorithm over random row stochastic matrices is shown. Then, the convergence of the proposed algorithm to the optimal cost is proven in expected value. Exact convergence is achieved when using diminishing (local) stepsizes, whereas approximate convergence is attained when constant stepsizes are employed. The convergence rate is shown to be sublinear and an explicit rate is provided in the case of constant stepsizes. Finally, the algorithm is tested on a distributed classification problem, first on synthetic data and, then, on a real, high-dimensional, text dataset.
引用
收藏
页码:4000 / 4014
页数:15
相关论文
共 51 条
[1]  
Agarwal Alekh, 2011, Advances in neural information processing systems, V24
[2]  
[Anonymous], 2003, STOCHASTIC APPROXIMA, DOI DOI 10.1007/978-1-4471-4285-0_3
[3]  
[Anonymous], 2014, Advances in Neural Information Processing Systems
[4]  
Bauschke H. H., 2001, Studies in Computational Mathematics, V8, P23, DOI DOI 10.1016/S1570-579X(01)80004-5
[5]   ON THE CONVERGENCE OF BLOCK COORDINATE DESCENT TYPE METHODS [J].
Beck, Amir ;
Tetruashvili, Luba .
SIAM JOURNAL ON OPTIMIZATION, 2013, 23 (04) :2037-2060
[6]   Large-Scale Machine Learning with Stochastic Gradient Descent [J].
Bottou, Leon .
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, :177-186
[7]  
Chen AI, 2012, ANN ALLERTON CONF, P601, DOI 10.1109/Allerton.2012.6483273
[8]  
Chen J., 2016, C UNC ART INT, P132
[9]   STOCHASTIC BLOCK MIRROR DESCENT METHODS FOR NONSMOOTH AND STOCHASTIC OPTIMIZATION [J].
Dang, Cong D. ;
Lan, Guanghui .
SIAM JOURNAL ON OPTIMIZATION, 2015, 25 (02) :856-881
[10]   Dual Averaging for Distributed Optimization: Convergence Analysis and Network Scaling [J].
Duchi, John C. ;
Agarwal, Alekh ;
Wainwright, Martin J. .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2012, 57 (03) :592-606