Stochastic variable metric proximal gradient with variance reduction for non-convex composite optimization

被引:0
作者
Gersende Fort
Eric Moulines
机构
[1] CNRS & Université de Toulouse,Institut de Mathématiques de Toulouse
[2] Ecole Polytechnique,CMAP
关键词
Stochastic optimization; Variable metric forward–backward splitting; Preconditioned stochastic gradient descent; Incremental expectation maximization; Proximal methods; Variance reduction; Non-asymptotic convergence bounds;
D O I
暂无
中图分类号
学科分类号
摘要
This paper introduces a novel algorithm, the Perturbed Proximal Preconditioned SPIDER algorithm (3P-SPIDER), designed to solve finite sum non-convex composite optimization. It is a stochastic Variable Metric Forward–Backward algorithm, which allows approximate preconditioned forward operator and uses a variable metric proximity operator as the backward operator; it also proposes a mini-batch strategy with variance reduction to address the finite sum setting. We show that 3P-SPIDER extends some Stochastic preconditioned Gradient Descent-based algorithms and some Incremental Expectation Maximization algorithms to composite optimization and to the case the forward operator can not be computed in closed form. We also provide an explicit control of convergence in expectation of 3P-SPIDER, and study its complexity in order to satisfy the approximate epsilon-stationary condition. Our results are the first to combine the non-convex composite optimization setting, a variance reduction technique to tackle the finite sum setting by using a minibatch strategy and, to allow deterministic or random approximations of the preconditioned forward operator. Finally, through an application to inference in a logistic regression model with random effects, we numerically compare 3P-SPIDER to other stochastic forward–backward algorithms and discuss the role of some design parameters of 3P-SPIDER.
引用
收藏
相关论文
共 77 条
[21]  
Choi H(2016)Mini-batch stochastic approximation methods for nonconvex stochastic composite optimization Math. Program. 4 634-299
[22]  
Hobert J(2022)Adaptivity of stochastic gradient methods for nonconvex optimization SIAM J. Math. Data Sci. 24 1420-55
[23]  
Chouzenoux E(2014)Proximal Newton-Type Methods for Minimizing Composite Functions SIAM J. Opt. 22 1-48
[24]  
Pesquet JC(2021)Stochastic proximal methods for non-smooth non-convex constrained sparse optimization J. Mach. Learn. Res. 93 273-1349
[25]  
Repetti A(1965)Proximité et dualité dans un espace hilbertien Bull. Soc. Math. France 13 45-1241
[26]  
Combettes P(2003)On the choice of the number of blocks with the incremental EM algorithm for the fitting of normal mixtures Stat. Comput. 21 1-704
[27]  
Vũ B(2020)ProxSARAH: an efficient algorithmic framework for stochastic composite nonconvex optimization J. Mach. Learn. Res. 108 1339-103
[28]  
Combettes PL(2013)Bayesian inference for logistic models using P’olya-Gamma latent variables J. Am. Stat. Assoc. 31 1215-4397
[29]  
Wajs VR(2021)Variable metric forward-backward algorithm for composite minimization problems SIAM J. Opt. 85 699-63
[30]  
Delyon B(1990)A Monte Carlo implementation of the EM algorithm and the poor man’s data augmentation algorithms J. Am. Stat. Assoc. 11 95-undefined