Run-and-Inspect Method for nonconvex optimization and global optimality bounds for R-local minimizers

被引:3
作者
Chen, Yifan [1 ]
Sun, Yuejiao [2 ]
Yin, Wotao [2 ]
机构
[1] Tsinghua Univ, Dept Math Sci, Beijing, Peoples R China
[2] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
关键词
R-local minimizer; Run-and-Inspect Method; Nonconvex optimization; Global minimum; Global optimality; CUBIC-REGULARIZATION; ALGORITHM;
D O I
10.1007/s10107-019-01397-w
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Many optimization algorithms converge to stationary points. When the underlying problem is nonconvex, they may get trapped at local minimizers or stagnate near saddle points. We propose the Run-and-Inspect Method, which adds an inspect phase to existing algorithms that helps escape from non-global stationary points. It samples a set of points in a radius R around the current point. When a sample point yields a sufficient decrease in the objective, we resume an existing algorithm from that point. If no sufficient decrease is found, the current point is called an approximate R-local minimizer. We show that an R-local minimizer is globally optimal, up to a specific error depending on R, if the objective function can be implicitly decomposed into a smooth convex function plus a restricted function that is possibly nonconvex, nonsmooth. Therefore, for such nonconvex objective functions, verifying global optimality is fundamentally easier. For high-dimensional problems, we introduce blockwise inspections to overcome the curse of dimensionality while still maintaining optimality bounds up to a factor equal to the number of blocks. We also present the sample complexities of these methods. When we apply our method to the existing algorithms on a set of artificial and realistic nonconvex problems, we find significantly improved chances of obtaining global minima.
引用
收藏
页码:39 / 67
页数:29
相关论文
共 29 条
[1]  
[Anonymous], 2017, ARXIV170300887
[2]  
[Anonymous], ARXIV160500405 CORR
[3]  
[Anonymous], 2017, ARXIV170610239
[4]  
[Anonymous], C LEARN THEOR
[5]  
Chaudhari P., 2016, ARXIV161101838
[6]   Deep relaxation: partial differential equations for optimizing deep neural networks [J].
Chaudhari, Pratik ;
Oberman, Adam ;
Osher, Stanley ;
Soatto, Stefano ;
Carlier, Guillaume .
RESEARCH IN THE MATHEMATICAL SCIENCES, 2018, 5 :1-30
[7]  
Conn AR, 2009, MOS-SIAM SER OPTIMIZ, V8, P1
[8]  
Conn Andrew R., 2000, SIAM, DOI DOI 10.1137/1.9780898719857
[9]  
Fox J., 2002, Cox Proportional-Hazads Regression for Survival Data
[10]  
Ge R, 2016, ADV NEUR IN, V29