Exploiting Hessian matrix and trust-region algorithm in hyperparameters estimation of Gaussian process

被引:38
作者
Zhang, YN [1 ]
Leithead, WE
机构
[1] Natl Univ Ireland, Hamilton Inst, Maynooth, Kildare, Ireland
[2] Univ Strathclyde, Dept Elect & Elect Engn, Glasgow G1 1QE, Lanark, Scotland
关键词
Gaussian process; log likelihood maximization; conjugate gradient; trust region; Hessian matrix;
D O I
10.1016/j.amc.2005.01.113
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Gaussian process (GP) regression is a Bayesian non-parametric regression model, showing good performance in various applications. However, it is quite rare to see research results on log-likelihood maximization algorithms. Instead of the commonly used conjugate gradient method, the Hessian matrix is first derived/simplified in this paper and the trust-region optimization method is then presented to estimate GP hyper-parameters. Numerical experiments verify the theoretical analysis, showing the advantages of using Hessian matrix and trust-region algorithms. In the GP context, the trust-region optimization method is a robust alternative to conjugate gradient method, also in view of future researches on approximate and/or parallel GP-implementation. (c) 2005 Elsevier Inc. All rights reserved.
引用
收藏
页码:1264 / 1281
页数:18
相关论文
共 28 条
[1]   A trust-region framework for managing the use of approximation models in optimization [J].
Alexandrov, NM ;
Dennis, JE ;
Lewis, RM ;
Torczon, V .
STRUCTURAL OPTIMIZATION, 1998, 15 (01) :16-23
[2]  
[Anonymous], EUR CONTR C CAMBR
[3]  
[Anonymous], ACTA NUMERICA 1992
[4]   What Size Net Gives Valid Generalization? [J].
Baum, Eric B. ;
Haussler, David .
NEURAL COMPUTATION, 1989, 1 (01) :151-160
[5]   APPROXIMATE SOLUTION OF THE TRUST REGION PROBLEM BY MINIMIZATION OVER TWO-DIMENSIONAL SUBSPACES [J].
BYRD, RH ;
SCHNABEL, RB ;
SHULTZ, GA .
MATHEMATICAL PROGRAMMING, 1988, 40 (03) :247-263
[6]  
Conn A., 2000, MOS-SIAM Series on Optimization
[7]  
Cressie N, 1993, STAT SPATIAL DATA
[8]  
Dennis, 1996, NUMERICAL METHODS UN
[9]  
Golub G.H., 2013, Matrix Computations, V4th
[10]   MULTILAYER FEEDFORWARD NETWORKS ARE UNIVERSAL APPROXIMATORS [J].
HORNIK, K ;
STINCHCOMBE, M ;
WHITE, H .
NEURAL NETWORKS, 1989, 2 (05) :359-366