A Bregman-Kaczmarz method for nonlinear systems of equations

被引:0
作者
Gower, Robert [1 ]
Lorenz, Dirk A. [2 ,3 ]
Winkler, Maximilian [2 ,3 ]
机构
[1] Simons Fdn, Flatiron Inst, CCM, New York, NY USA
[2] TU Braunschweig, Inst Anal & Algebra, Braunschweig, Germany
[3] Univ Bremen, Ctr Ind Math, Fachbereich 3, Bremen, Germany
关键词
Nonlinear systems; Stochastic methods; Randomized Kaczmarz; Bregman projections; FEASIBILITY PROBLEMS; PROJECTION METHODS; CONVERGENCE; RECONSTRUCTION; REGULARIZATION; ALGORITHMS;
D O I
10.1007/s10589-023-00541-9
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
We propose a new randomized method for solving systems of nonlinear equations, which can find sparse solutions or solutions under certain simple constraints. The scheme only takes gradients of component functions and uses Bregman projections onto the solution space of a Newton equation. In the special case of euclidean projections, the method is known as nonlinear Kaczmarz method. Furthermore if the component functions are nonnegative, we are in the setting of optimization under the interpolation assumption and the method reduces to SGD with the recently proposed stochastic Polyak step size. For general Bregman projections, our method is a stochastic mirror descent with a novel adaptive step size. We prove that in the convex setting each iteration of our method results in a smaller Bregman distance to exact solutions as compared to the standard Polyak step. Our generalization to Bregman projections comes with the price that a convex one-dimensional optimization problem needs to be solved in each iteration. This can typically be done with globalized Newton iterations. Convergence is proved in two classical settings of nonlinearity: for convex nonnegative functions and locally for functions which fulfill the tangential cone condition. Finally, we show examples in which the proposed method outperforms similar methods with the same memory requirements.
引用
收藏
页码:1059 / 1098
页数:40
相关论文
共 65 条
[1]   Convergence of Bregman projection methods for solving consistent convex feasibility problems in reflexive Banach spaces [J].
Alber, Y ;
Butnariu, D .
JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 1997, 92 (01) :33-61
[2]  
[Anonymous], 2001, Studies in Computational Mathematics, DOI 10.1016/S1570-579X(01)80009-4
[3]  
Arora R, 2013, J MACH LEARN RES, V14, P1715
[4]   Stochastic Mirror Descent on Overparameterized Nonlinear Models [J].
Azizan, Navid ;
Lale, Sahin ;
Hassibi, Babak .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (12) :7717-7727
[5]  
Bauschke H. H., 1997, Journal of Convex Analysis, V4, P27
[6]  
Bauschke HH, 2011, CMS BOOKS MATH, P1, DOI 10.1007/978-1-4419-9467-7
[7]   Bregman monotone optimization algorithms [J].
Bauschke, HH ;
Borwein, JM ;
Combettes, PL .
SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2003, 42 (02) :596-636
[8]   Iterating Bregman retractions [J].
Bauschke, HH ;
Combettes, AL .
SIAM JOURNAL ON OPTIMIZATION, 2003, 13 (04) :1159-1173
[9]  
Beck A, 2017, MOS-SIAM SER OPTIMIZ, P1, DOI 10.1137/1.9781611974997
[10]   ITERATIVE BREGMAN PROJECTIONS FOR REGULARIZED TRANSPORTATION PROBLEMS [J].
Benamou, Jean-David ;
Carlier, Guillaume ;
Cuturi, Marco ;
Nenna, Luca ;
Peyre, Gabriel .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2015, 37 (02) :A1111-A1138