QR decomposition based low rank approximation for Gaussian process regression

被引:0
作者
Emil Thomas
Vivek Sarin
机构
[1] Texas A &M University,Department of Computer Science & Engineering
来源
Applied Intelligence | 2023年 / 53卷
关键词
Gaussian process; Nyström approximation; QR decomposition; Low-rank approximation;
D O I
暂无
中图分类号
学科分类号
摘要
This paper presents a QR decomposition based low-rank approximation algorithm for training and prediction in Gaussian process regression. Conventional low-rank methods like FITC, Nyström, etc., rely on low-rank approximations to the full kernel matrix that are derived from a set of representative data points. Training with FITC involves the selection of these representative points and is computationally intensive when the dimension of the dataset is large. Prediction accuracy of Nyström suffers when the number of representative points is small or when the length scale is small. The algorithm proposed here uses the thin QR decomposition of the low-rank matrix used in the Nyström approximation. We show that the marginal likelihood and its derivatives needed for training can be split into the subspace belonging to that approximation and its orthogonal complement. During prediction, we further restrict the target vectors into this subspace and thus eliminate the numerical error caused by very low noise variance. Use of the QR factorization improves the training and prediction performance. Experiments on real and synthetic data show that our approach is more accurate and faster than conventional methods. Our algorithm performs well at low length scales and achieves the prediction accuracy comparable to the full kernel matrix approach even for low rank approximations, without the need to optimize the representative points. This results in a simpler and faster approximation that could be used to scale Gaussian process regression to large datasets.
引用
收藏
页码:28924 / 28936
页数:12
相关论文
共 31 条
[1]  
Chalupka K(2013)A framework for evaluating approximation methods for gaussian process regression J Mach Learn Res 14 333-350
[2]  
Williams CK(2011)Libsvm: A library for support vector machines ACM Trans Intell Syst Technol 2 1-27
[3]  
Murray I(1981)Bayes empirical bayes J Am Stat Assoc 76 833-841
[4]  
Chang CC(2018)Patchwork kriging for large-scale gaussian process regression J Mach Learn Res 19 269-311
[5]  
Lin CJ(2012)Gplp: a local and parallel computation toolbox for gaussian process regression J Mach Learn Res 13 775-779
[6]  
Deely J(2019)Improved fixed-rank Nyström approximation via qr decomposition: Practical and theoretical aspects Neurocomputing 363 261-272
[7]  
Lindley D(2010)Gaussian processes for machine learning (gpml) toolbox J Mach Learn Res 11 3011-3015
[8]  
Park C(2006)Sparse gaussian processes using pseudo-inputs Adv Neural Inf Process Syst 18 1259-1266
[9]  
Apley D(1999)Least squares support vector machine classifiers Neural Process Lett 9 293-300
[10]  
Park C(2022)Sparse gaussian process regression for multi-step ahead forecasting of wind gusts combining numerical weather predictions and on-site measurements J Wind Eng Ind Aerodyn 220 5260-5275