CONVERGENCE OF TRUST-REGION METHODS BASED ON PROBABILISTIC MODELS

被引:74
作者
Bandeira, A. S. [1 ]
Scheinberg, K. [2 ]
Vicente, L. N. [3 ]
机构
[1] Princeton Univ, Program Appl & Computat Math, Princeton, NJ 08544 USA
[2] Lehigh Univ, Dept Ind & Syst Engn, Bethlehem, PA 18015 USA
[3] Univ Coimbra, Dept Math, CMUC, P-3001454 Coimbra, Portugal
基金
美国国家科学基金会;
关键词
trust-region methods; unconstrained optimization; probabilistic models; derivative-free optimization; global convergence; DERIVATIVE-FREE OPTIMIZATION; CONDITION NUMBERS; QUADRATIC MODELS; DIRECT SEARCH; ALGORITHMS; GEOMETRY; SETS;
D O I
10.1137/130915984
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
In this paper we consider the use of probabilistic or random models within a classical trust-region framework for optimization of deterministic smooth general nonlinear functions. Our method and setting differs from many stochastic optimization approaches in two principal ways. Firstly, we assume that the value of the function itself can be computed without noise, in other words, that the function is deterministic. Second, we use random models of higher quality than those produced by the usual stochastic gradient methods. In particular, a first order model based on random approximation of the gradient is required to provide sufficient quality of approximation with probability >= 1/2. This is in contrast with stochastic gradient approaches, where the model is assumed to be "correct" only in expectation. As a result of this particular setting, we are able to prove convergence, with probability one, of a trust-region method which is almost identical to the classical method. Moreover, the new method is simpler than its deterministic counterpart as it does not require a criticality step. Hence we show that a standard optimization framework can be used in cases when models are random and may or may not provide good approximations, as long as "good" models are more likely than "bad" models. Our results are based on the use of properties of martingales. Our motivation comes from using random sample sets and interpolation models in derivative-free optimization. However, our framework is general and can be applied with any source of uncertainty in the model. We discuss various applications for our methods in the paper.
引用
收藏
页码:1238 / 1264
页数:27
相关论文
共 35 条
[1]  
[Anonymous], 2008, 10 COPP MOUNT C IT M
[2]  
[Anonymous], 2010, Theoretical foundations and numerical methods for sparse recovery, DOI DOI 10.1515/9783110226157.1
[3]   Mesh adaptive direct search algorithms for constrained optimization [J].
Audet, C ;
Dennis, JE .
SIAM JOURNAL ON OPTIMIZATION, 2006, 17 (01) :188-217
[4]  
Ayaz U, 2014, ELECTRON T NUMER ANA, V41, P167
[5]   Computation of sparse low degree interpolating polynomials and their application to derivative-free optimization [J].
Bandeira, A. S. ;
Scheinberg, K. ;
Vicente, L. N. .
MATHEMATICAL PROGRAMMING, 2012, 134 (01) :223-257
[6]   DERIVATIVE-FREE OPTIMIZATION OF EXPENSIVE FUNCTIONS WITH COMPUTATIONAL ERROR USING WEIGHTED REGRESSION [J].
Billups, Stephen C. ;
Larson, Jeffrey ;
Graf, Peter .
SIAM JOURNAL ON OPTIMIZATION, 2013, 23 (01) :27-53
[7]   Sample size selection in optimization methods for machine learning [J].
Byrd, Richard H. ;
Chin, Gillian M. ;
Nocedal, Jorge ;
Wu, Yuchen .
MATHEMATICAL PROGRAMMING, 2012, 134 (01) :127-155
[8]  
Candes E. J., 2006, P INT C MATH MADR SP, V3, P1433, DOI DOI 10.4171/022-3/69
[9]   Multiperiod consumption and portfolio decisions under the multivariate GARCH model with transaction costs and CVaR-based risk control [J].
Chen, ZP .
OR SPECTRUM, 2005, 27 (04) :603-632
[10]  
Conn A, 1997, On the convergence of derivative-free methods for unconstrained optimization, P83