Markov chain block coordinate descent

被引:0
作者
Tao Sun
Yuejiao Sun
Yangyang Xu
Wotao Yin
机构
[1] National University of Defense Technology,National Lab for Parallel and Distributed Processing, College of Computer
[2] University of California,Department of Mathematics
[3] Rensselaer Polytechnic Institute,Department of Mathematical Sciences
来源
Computational Optimization and Applications | 2020年 / 75卷
关键词
Block coordinate gradient descent; Markov chain; Markov chain Monte Carlo; Markov decision process; Decentralized optimization; Primary 90C26; Secondary 90C40; 68W15;
D O I
暂无
中图分类号
学科分类号
摘要
The method of block coordinate gradient descent (BCD) has been a powerful method for large-scale optimization. This paper considers the BCD method that successively updates a series of blocks selected according to a Markov chain. This kind of block selection is neither i.i.d. random nor cyclic. On the other hand, it is a natural choice for some applications in distributed optimization and Markov decision process, where i.i.d. random and cyclic selections are either infeasible or very expensive. By applying mixing-time properties of a Markov chain, we prove convergence of Markov chain BCD for minimizing Lipschitz differentiable functions, which can be nonconvex. When the functions are convex and strongly convex, we establish both sublinear and linear convergence rates, respectively. We also present a method of Markov chain inertial BCD. Finally, we discuss potential applications.
引用
收藏
页码:35 / 61
页数:26
相关论文
共 51 条
[1]  
Beck A(2013)On the convergence of block coordinate descent type methods SIAM J. Optim. 23 2037-2060
[2]  
Tetruashvili L(2005)Basic properties of strong mixing conditions—a survey and some open questions Probab Surv 2 107-144
[3]  
Bradley RC(1999)Resource-constrained project scheduling: notation, classification, models, and methods Eur. J. Oper. Res. 112 3-41
[4]  
Brucker P(2017)Cyclic coordinate-update algorithms for fixed-point problems: analysis and applications SIAM J. Sci. Comput. 39 A1280-A1300
[5]  
Drexl A(2015)Stochastic block mirror descent methods for nonsmooth and stochastic optimization SIAM J. Optim. 25 856-881
[6]  
Möhring R(2015)Accelerated, parallel, and proximal coordinate descent SIAM J. Optim. 25 1997-2023
[7]  
Neumann K(2009)Coordinate descent optimization for Inverse Probl. Imaging 3 487-503
[8]  
Pesch E(2015) minimization with application to compressed sensing; a greedy algorithm SIAM J. Optim. 25 210-233
[9]  
Chow YT(2015)On convergence of the maximum block improvement method SIAM J. Optim. 25 351-376
[10]  
Wu T(2012)Asynchronous stochastic coordinate descent: parallelism and convergence properties SIAM J. Optim. 22 341-362