The Optimal Ridge Penalty for Real-world High-dimensional Data Can Be Zero or Negative due to the Implicit Ridge Regularization

被引:0
|
作者
Kobak, Dmitry [1 ]
Lomond, Jonathan [1 ]
Sanchez, Benoit [1 ]
机构
[1] Univ Tubingen, Inst Ophthalm Res, Otfried Muller Str 25, D-72076 Tubingen, Germany
基金
美国国家卫生研究院;
关键词
High-dimensional; ridge regression; regularization; REGRESSION; SELECTION;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A conventional wisdom in statistical learning is that large models require strong regularization to prevent overfitting. Here we show that this rule can be violated by linear regression in the underdetermined n << p situation under realistic conditions. Using simulations and real-life high-dimensional datasets, we demonstrate that an explicit positive ridge penalty can fail to provide any improvement over the minimum-norm least squares estimator. Moreover, the optimal value of ridge penalty in this situation can be negative. This happens when the high-variance directions in the predictor space can predict the response variable, which is often the case in the real-world high-dimensional data. In this regime, low-variance directions provide an implicit ridge regularization and can make any further positive ridge penalty detrimental. We prove that augmenting any linear model with random covariates and using minimum-norm estimator is asymptotically equivalent to adding the ridge penalty. We use a spiked covariance model as an analytically tractable example and prove that the optimal ridge penalty in this case is negative when n << p.
引用
收藏
页数:16
相关论文
共 10 条
  • [1] Ridge estimation of inverse covariance matrices from high-dimensional data
    van Wieringen, Wessel N.
    Peeters, Carel F. W.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2016, 103 : 284 - 303
  • [2] Robust Ridge Regression for High-Dimensional Data
    Maronna, Ricardo A.
    TECHNOMETRICS, 2011, 53 (01) : 44 - 53
  • [3] Discriminative Ridge Machine: A Classifier for High-Dimensional Data or Imbalanced Data
    Peng, Chong
    Cheng, Qiang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (06) : 2595 - 2609
  • [4] Estimation of variance components, heritability and the ridge penalty in high-dimensional generalized linear models
    Veerman, Jurre R.
    Leday, Gwenael G. R.
    van de Wiel, Mark A.
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2022, 51 (01) : 116 - 134
  • [5] Non-negative Constrained Penalty for High-Dimensional Correlated Data
    Ming, Hao
    Chen, Yinjun
    Yang, Hu
    COMMUNICATIONS IN MATHEMATICS AND STATISTICS, 2025,
  • [6] Enhancing Parameters Tuning of Overlay Models with Ridge Regression: Addressing Multicollinearity in High-Dimensional Data
    Magklaras, Aris
    Gogos, Christos
    Alefragis, Panayiotis
    Birbas, Alexios
    MATHEMATICS, 2024, 12 (20)
  • [7] Machine learning methods for propensity and disease risk score estimation in high-dimensional data: a plasmode simulation and real-world data cohort analysis
    Guo, Yuchen
    Strauss, Victoria Y.
    Catala, Marti
    Jodicke, Annika M.
    Khalid, Sara
    Prieto-Alhambra, Daniel
    FRONTIERS IN PHARMACOLOGY, 2024, 15
  • [8] Targeted Fused Ridge Estimation of Inverse Covariance Matrices from Multiple High-Dimensional Data Classes
    Bilgrau, Anders Ellern
    Peeters, Carel F. W.
    Eriksen, Poul Svante
    Bogsted, Martin
    van Wieringen, Wessel N.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [9] Feature Selection for High-Dimensional Data Based on Ridge Regression and SVM and Its Application in Peptide QSAR Modeling
    Wang Zhi-Ming
    Han Na
    Yuan Zhe-Ming
    Wu Zhao-Hua
    ACTA PHYSICO-CHIMICA SINICA, 2013, 29 (03) : 498 - 507
  • [10] An Instance-Frequency-Weighted Regularization Scheme for Non-Negative Latent Factor Analysis on High-Dimensional and Sparse Data
    Luo, Xin
    Wang, Zidong
    Shang, Mingsheng
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2021, 51 (06): : 3522 - 3532