Scalable subspace methods for derivative-free nonlinear least-squares optimization

被引:0
作者
Coralia Cartis
Lindon Roberts
机构
[1] University of Oxford,Mathematical Institute
[2] Australian National University,Mathematical Sciences Institute, Building 145, Science Road
来源
Mathematical Programming | 2023年 / 199卷
关键词
Derivative-free optimization; Large-scale optimization; Nonlinear least-squares; Worst case complexity; 65K05; 90C30; 90C56;
D O I
暂无
中图分类号
学科分类号
摘要
We introduce a general framework for large-scale model-based derivative-free optimization based on iterative minimization within random subspaces. We present a probabilistic worst-case complexity analysis for our method, where in particular we prove high-probability bounds on the number of iterations before a given optimality is achieved. This framework is specialized to nonlinear least-squares problems, with a model-based framework based on the Gauss–Newton method. This method achieves scalability by constructing local linear interpolation models to approximate the Jacobian, and computes new steps at each iteration in a subspace with user-determined dimension. We then describe a practical implementation of this framework, which we call DFBGN. We outline efficient techniques for selecting the interpolation points and search subspace, yielding an implementation that has a low per-iteration linear algebra cost (linear in the problem dimension) while also achieving fast objective decrease as measured by evaluations. Extensive numerical results demonstrate that DFBGN has improved scalability, yielding strong performance on large-scale nonlinear least-squares problems.
引用
收藏
页码:461 / 524
页数:63
相关论文
共 144 条
  • [1] Alarie S(2021)Two decades of blackbox optimization applications EURO J. Comput. Optim. 9 100011-257
  • [2] Audet C(2012)Computation of sparse low degree interpolating polynomials and their application to derivative-free optimization Math. Program. 134 223-1264
  • [3] Gheribi AE(2014)Convergence of trust-region methods based on probabilistic models SIAM J. Optim. 24 1238-2506
  • [4] Kokkolaras M(2016)Sharp nonasymptotic bounds on the norm of random matrices with independent entries Ann. Probab. 44 2479-680
  • [5] Le Digabel S(2020)An investigation of Newton–Sketch and subsampled Newton methods Optim. Methods Softw. 35 661-560
  • [6] Bandeira AS(2022)A theoretical and empirical comparison of gradient approximations in derivative-free optimization Found. Comput. Math. 22 507-951
  • [7] Scheinberg K(2016)Levenberg–Marquardt methods based on probabilistic gradient models and inexact subproblem solution, with application to data assimilation SIAM/ASA J. Uncertain. Quantif. 4 924-2749
  • [8] Vicente LN(2020)Stochastic three points method for unconstrained smooth minimization SIAM J. Optim. 30 2726-119
  • [9] Bandeira AS(2019)Convergence rate analysis of a stochastic trust region method for nonconvex optimization INFORMS J. Optim. 1 92-32:41
  • [10] Scheinberg K(2019)Improving the flexibility and robustness of model-based derivative-free optimization solvers ACM Trans. Math. Softw. 45 32:1-674