Stable prediction in high-dimensional linear models

被引:18
|
作者
Lin, Bingqing [1 ]
Wang, Qihua [1 ,2 ]
Zhang, Jun [1 ]
Pang, Zhen [3 ]
机构
[1] Shenzhen Univ, Inst Stat Sci, Coll Math & Stat, Shenzhen 518060, Peoples R China
[2] Chinese Acad Sci, Acad Math & Syst Sci, Beijing 100190, Peoples R China
[3] Hong Kong Polytech Univ, Dept Appl Math, Kowloon, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Model averaging; Variable selection; Penalized regression; Screening; VARIABLE SELECTION; REGRESSION;
D O I
10.1007/s11222-016-9694-6
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We propose a Random Splitting Model Averaging procedure, RSMA, to achieve stable predictions in high-dimensional linear models. The idea is to use split training data to construct and estimate candidate models and use test data to form a second-level data. The second-level data is used to estimate optimal weights for candidate models by quadratic optimization under non-negative constraints. This procedure has three appealing features: (1) RSMA avoids model overfitting, as a result, gives improved prediction accuracy. (2) By adaptively choosing optimal weights, we obtain more stable predictions via averaging over several candidate models. (3) Based on RSMA, a weighted importance index is proposed to rank the predictors to discriminate relevant predictors from irrelevant ones. Simulation studies and a real data analysis demonstrate that RSMA procedure has excellent predictive performance and the associated weighted importance index could well rank the predictors.
引用
收藏
页码:1401 / 1412
页数:12
相关论文
共 50 条
  • [21] Learning High-Dimensional Generalized Linear Autoregressive Models
    Hall, Eric C.
    Raskutti, Garvesh
    Willett, Rebecca M.
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2019, 65 (04) : 2401 - 2422
  • [22] TESTING HIGH-DIMENSIONAL REGRESSION COEFFICIENTS IN LINEAR MODELS
    Zhao, Alex
    Li, Changcheng
    Li, Runze
    Zhang, Zhe
    ANNALS OF STATISTICS, 2024, 52 (05): : 2034 - 2058
  • [23] TESTABILITY OF HIGH-DIMENSIONAL LINEAR MODELS WITH NONSPARSE STRUCTURES
    Bradic, Jelena
    Fan, Jianqing
    Zhu, Yinchu
    ANNALS OF STATISTICS, 2022, 50 (02): : 615 - 639
  • [24] High-dimensional robust inference for censored linear models
    Huang, Jiayu
    Wu, Yuanshan
    SCIENCE CHINA-MATHEMATICS, 2024, 67 (04) : 891 - 918
  • [25] Noise Level Estimation in High-Dimensional Linear Models
    Golubev, G. K.
    Krymova, E. A.
    PROBLEMS OF INFORMATION TRANSMISSION, 2018, 54 (04) : 351 - 371
  • [26] Tests for high-dimensional partially linear regression models
    Shi, Hongwei
    Yang, Weichao
    Sun, Bowen
    Guo, Xu
    STATISTICAL PAPERS, 2025, 66 (03)
  • [27] Estimation in High-Dimensional Analysis and Multivariate Linear Models
    Kollo, Tonu
    Von Rosen, Tatjana
    Von Rosen, Dietrich
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 2011, 40 (07) : 1241 - 1253
  • [28] Empirical likelihood for high-dimensional linear regression models
    Hong Guo
    Changliang Zou
    Zhaojun Wang
    Bin Chen
    Metrika, 2014, 77 : 921 - 945
  • [29] Spatially relaxed inference on high-dimensional linear models
    Jérôme-Alexis Chevalier
    Tuan-Binh Nguyen
    Bertrand Thirion
    Joseph Salmon
    Statistics and Computing, 2022, 32
  • [30] High-Dimensional Linear Models: A Random Matrix Perspective
    Jamshid Namdari
    Debashis Paul
    Lili Wang
    Sankhya A, 2021, 83 (2): : 645 - 695