Risk Convergence of Centered Kernel Ridge Regression With Large Dimensional Data

被引:8
作者
Elkhalil, Khalil [1 ]
Kammoun, Abla [1 ]
Zhang, Xiangliang [1 ]
Alouini, Mohamed-Slim [1 ]
Al-Naffouri, Tareq [1 ]
机构
[1] King Abdullah Univ Sci & Technol, Elect Engn Program, Thuwal 23955, Saudi Arabia
关键词
Kernel; Training; Convergence; Training data; Aerospace electronics; Optimization; Predictive models; Kernel regression; centered kernels; random matrix theory;
D O I
10.1109/TSP.2020.2975939
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper carries out a large dimensional analysis of a variation of kernel ridge regression that we call centered kernel ridge regression (CKRR), also known in the literature as kernel ridge regression with offset. This modified technique is obtained by accounting for the bias in the regression problem resulting in the old kernel ridge regression but with centered kernels. The analysis is carried out under the assumption that the data is drawn from a Gaussian distribution and heavily relies on tools from random matrix theory (RMT). Under the regime in which the data dimension and the training size grow infinitely large with fixed ratio and under some mild assumptions controlling the data statistics, we show that both the empirical and the prediction risks converge to a deterministic quantities that describe in closed form fashion the performance ofCKRRin terms of the data statistics and dimensions. Inspired by this theoretical result, we subsequently build a consistent estimator of the prediction risk based on the training data which allows to optimally tune the design parameters. Akey insight of the proposed analysis is the fact that asymptotically a large class of kernels achieve the same minimum prediction risk. This insight is validated with both synthetic and real data.
引用
收藏
页码:1574 / 1588
页数:15
相关论文
共 23 条
  • [1] Alaoui A., 2015, Advances in Neural Information Processing Systems, V28
  • [2] [Anonymous], RANDOM MATRICES THEO
  • [3] [Anonymous], P ADV NEURAL INF PRO
  • [4] Bai Z, 2010, SPRINGER SER STAT, P1, DOI 10.1007/978-1-4419-0661-8
  • [5] Bishop Christopher M, 2006, MACH LEARN, V128, DOI DOI 10.1117/1.2819119
  • [6] Caponnetto A, 2007, FOUND COMPUT MATH, V7, P331, DOI [10.1007/s10208-006-0196-8, 10.1007/S10208-006-0196-8]
  • [7] Centered Kernel Alignment Enhancing Neural Network Pretraining for MRI-Based Dementia Diagnosis
    Cardenas-Pena, David
    Collazos-Huertas, Diego
    Castellanos-Dominguez, German
    [J]. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2016, 2016
  • [8] Model complexity control for regression using VC generalization bounds
    Cherkassky, V
    Shao, XH
    Mulier, FM
    Vapnik, VN
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1999, 10 (05): : 1075 - 1089
  • [9] Cortes C, 2012, J MACH LEARN RES, V13, P795
  • [10] Couillet R., 2011, RANDOM MATRIX METHOD