Adaptive Restart of the Optimized Gradient Method for Convex Optimization

被引:0
作者
Donghwan Kim
Jeffrey A. Fessler
机构
[1] University of Michigan,
来源
Journal of Optimization Theory and Applications | 2018年 / 178卷
关键词
Convex optimization; First-order methods; Accelerated gradient methods; Optimized gradient method; Restarting; 80M50; 90C06; 90C25;
D O I
暂无
中图分类号
学科分类号
摘要
First-order methods with momentum, such as Nesterov’s fast gradient method, are very useful for convex optimization problems, but can exhibit undesirable oscillations yielding slow convergence rates for some applications. An adaptive restarting scheme can improve the convergence rate of the fast gradient method, when the parameter of a strongly convex cost function is unknown or when the iterates of the algorithm enter a locally strongly convex region. Recently, we introduced the optimized gradient method, a first-order algorithm that has an inexpensive per-iteration computational cost similar to that of the fast gradient method, yet has a worst-case cost function rate that is twice faster than that of the fast gradient method and that is optimal for large-dimensional smooth convex problems. Building upon the success of accelerating the fast gradient method using adaptive restart, this paper investigates similar heuristic acceleration of the optimized gradient method. We first derive a new first-order method that resembles the optimized gradient method for strongly convex quadratic problems with known function parameters, yielding a linear convergence rate that is faster than that of the analogous version of the fast gradient method. We then provide a heuristic analysis and numerical experiments that illustrate that adaptive restart can accelerate the convergence of the optimized gradient method. Numerical results also illustrate that adaptive restart is helpful for a proximal version of the optimized gradient method for nonsmooth composite convex functions.
引用
收藏
页码:240 / 263
页数:23
相关论文
共 34 条
[1]  
Cevher V(2014)Convex optimization for big data: scalable, randomized, and parallel algorithms for big data analytics IEEE Signal Process. Mag. 31 32-43
[2]  
Becker S(1983)A method for unconstrained convex minimization problem with the rate of convergence Dokl. Akad. Nauk. USSR 269 543-7
[3]  
Schmidt M(2009)A fast iterative shrinkage-thresholding algorithm for linear inverse problems SIAM J. Imaging Sci. 2 183-202
[4]  
Nesterov Y(2015)Adaptive restart for accelerated gradient schemes Found. Comput. Math. 15 715-32
[5]  
Beck A(2016)A differential equation for modeling Nesterov’s accelerated gradient method: theory and insights J. Mach. Learn. Res. 17 1-43
[6]  
Teboulle M(2015)Fast parallel MR image reconstruction via B1-based, adaptive restart, iterative soft thresholding algorithms (BARISTA) IEEE Trans. Med. Imaging 34 578-88
[7]  
O’Donoghue B(2016)An adaptive accelerated first-order method for convex optimization Comput. Optim. Appl. 64 31-73
[8]  
Candès E(2016)Optimized first-order methods for smooth convex minimization Math. Program. 159 81-107
[9]  
Su W(2014)Performance of first-order methods for smooth convex minimization: a novel approach Math. Program. 145 451-82
[10]  
Boyd S(2017)The exact information-based complexity of smooth convex minimization J. Complex. 39 1-16