ROBUST ACCELERATED PRIMAL-DUAL METHODS FOR COMPUTING SADDLE POINTS

被引:0
作者
Zhang, Xuan [1 ]
Aybat, Necdet serhat [1 ]
Guerbuezbalaban, Mert [2 ]
机构
[1] Penn State Univ, Dept Ind & Mfg Engn, University Pk, PA 16802 USA
[2] Rutgers State Univ, Dept Management Sci & Informat Syst, Piscataway, NJ 08854 USA
关键词
saddle point problems; convex optimization; stochastic gradient; accelerated methods; VARIATIONAL-INEQUALITIES; GRADIENT METHODS; CONVEX;
D O I
10.1137/21M1462775
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider strongly-convex-strongly-concave saddle point problems assuming we have access to unbiased stochastic estimates of the gradients. We propose a stochastic accelerated primal-dual (SAPD) algorithm and show that the SAPD sequence, generated using constant primaldual step sizes, linearly converges to a neighborhood of the unique saddle point. Interpreting the size of the neighborhood as a measure of robustness to gradient noise, we obtain explicit characterizations of robustness in terms of SAPD parameters and problem constants. Based on these characterizations, we develop computationally tractable techniques for optimizing the SAPD parameters, i.e., the primal and dual step sizes, and the momentum parameter, to achieve a desired trade-off between the convergence rate and robustness on the Pareto curve. This allows SAPD to enjoy fast convergence properties while being robust to noise as an accelerated method. SAPD admits convergence guarantees for the distance metric with a variance term optimal up to a logarithmic factor, which can be removed by employing a restarting strategy. We also discuss how convergence and robustness results extend to the merely-convex-merely-concave setting. Finally, we illustrate our framework on a distributionally robust logistic regression problem.
引用
收藏
页码:1097 / 1130
页数:34
相关论文
共 53 条
[1]  
[Anonymous], 2004, Advances in neural information processing systems (NIPS)
[2]  
Aybat NS, 2019, ADV NEUR IN, V32
[3]   ROBUST ACCELERATED GRADIENT METHODS FOR SMOOTH STRONGLY CONVEX FUNCTIONS [J].
Aybat, Necdet Serhat ;
Fallah, Alireza ;
Gurbuzbalaban, Mert ;
Ozdaglar, Asuman .
SIAM JOURNAL ON OPTIMIZATION, 2020, 30 (01) :717-751
[4]  
Balamurugan P, 2016, ADV NEUR IN, V29
[5]  
Beck A, 2017, MOS-SIAM SER OPTIMIZ, P1, DOI 10.1137/1.9781611974997
[6]  
BenTal A, 2009, PRINC SER APPL MATH, P1
[7]   Optimization Methods for Large-Scale Machine Learning [J].
Bottou, Leon ;
Curtis, Frank E. ;
Nocedal, Jorge .
SIAM REVIEW, 2018, 60 (02) :223-311
[8]  
Can B, 2022, Arxiv, DOI arXiv:2202.09688
[9]   On the ergodic convergence rates of a first-order primal-dual algorithm [J].
Chambolle, Antonin ;
Pock, Thomas .
MATHEMATICAL PROGRAMMING, 2016, 159 (1-2) :253-287
[10]   A First-Order Primal-Dual Algorithm for Convex Problems with Applications to Imaging [J].
Chambolle, Antonin ;
Pock, Thomas .
JOURNAL OF MATHEMATICAL IMAGING AND VISION, 2011, 40 (01) :120-145