ADAPTIVE REGULARIZATION ALGORITHMS WITH INEXACT EVALUATIONS FOR NONCONVEX OPTIMIZATION

被引:27
作者
Bellavia, Stefania [1 ]
Guriol, Gianmarco [2 ]
Morini, Benedetta [1 ]
Toint, Philippe L. [3 ]
机构
[1] Univ Firenze, Dipartimento Ingn Ind, I-50134 Florence, Italy
[2] Univ Firenze, Dipartimento Matemat & Informat Ulisse Dini, I-50134 Florence, Italy
[3] Univ Namur, Namur Ctr Complex Syst naXys, 61 Rue Bruxelles, B-5000 Namur, Belgium
关键词
evaluation complexity; regularization methods; inexact functions and derivatives; subsampling methods; TRUST-REGION METHODS; COMPLEXITY;
D O I
10.1137/18M1226282
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
A regularization algorithm using inexact function values and inexact derivatives is proposed and its evaluation complexity analyzed. This algorithm is applicable to unconstrained problems and to problems with inexpensive constraints (that is, constraints whose evaluation and enforcement has negligible cost) under the assumption that the derivative of highest degree is beta-Holder continuous. It features a very flexible adaptive mechanism for determining the inexactness which is allowed, at each iteration, when computing objective function values and derivatives. The complexity analysis covers arbitrary optimality order and arbitrary degree of available approximate derivatives. It extends results of Cartis, Gould, and Toint [SIAM J. Optim., to appear] on the evaluation complexity to the inexact case: if a qth-order minimizer is sought using approximations to the first p derivatives, it is proved that a suitable approximate minimizer within epsilon is computed by the proposed algorithm in at most O(epsilon(-p+beta/p-q+beta)) iterations and at most O(vertical bar log(epsilon)vertical bar epsilon(-p+beta/p-q+beta)) approximate evaluations. An algorithmic variant, although more rigid in practice, can be proved to find such an approximate minimizer in O(vertical bar log(epsilon)vertical bar + epsilon(-p+beta/p-q+beta)) evaluations. While the proposed framework remains so far conceptual for high degrees and orders, it is shown to yield simple and computationally realistic inexact methods when specialized to the unconstrained and bound-constrained first-and second-order cases. The deterministic complexity results are finally extended to the stochastic context, yielding adaptive sample-size rules for subsampling methods typical of machine learning.
引用
收藏
页码:2881 / 2915
页数:35
相关论文
共 31 条
[1]  
[Anonymous], 2004, INTRO LECT CONVEX OP
[2]   CONVERGENCE OF TRUST-REGION METHODS BASED ON PROBABILISTIC MODELS [J].
Bandeira, A. S. ;
Scheinberg, K. ;
Vicente, L. N. .
SIAM JOURNAL ON OPTIMIZATION, 2014, 24 (03) :1238-1264
[3]  
BELLAVIA S., 2018, PREPRINT
[4]   A Levenberg-Marquardt method for large nonlinear least-squares problems with dynamic accuracy in functions and gradients [J].
Bellavia, Stefania ;
Gratton, Serge ;
Riccietti, Elisa .
NUMERISCHE MATHEMATIK, 2018, 140 (03) :791-825
[5]  
BERGOU E., 2018, PREPRINT
[6]   Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models [J].
Birgin, E. G. ;
Gardenghi, J. L. ;
Martinez, J. M. ;
Santos, S. A. ;
Toint, Ph. L. .
MATHEMATICAL PROGRAMMING, 2017, 163 (1-2) :359-368
[7]  
Blanchet J., 2019, INFORMS J OPTIM, V1, P92, DOI DOI 10.1287/IJOO.2019.0016
[8]  
BONNIOT T., 2018, THESIS
[9]   Optimization Methods for Large-Scale Machine Learning [J].
Bottou, Leon ;
Curtis, Frank E. ;
Nocedal, Jorge .
SIAM REVIEW, 2018, 60 (02) :223-311
[10]   Global convergence rate analysis of unconstrained optimization methods based on probabilistic models [J].
Cartis, C. ;
Scheinberg, K. .
MATHEMATICAL PROGRAMMING, 2018, 169 (02) :337-375