SPReM: Sparse Projection Regression Model For High-Dimensional Linear Regression

被引:11
|
作者
Sun, Qiang [1 ]
Zhu, Hongtu [2 ]
Liu, Yufeng [3 ]
Ibrahim, Joseph G. [2 ]
机构
[1] Univ N Carolina, Dept Biostat, Chapel Hill, NC 27599 USA
[2] Univ N Carolina, Dept Biostat, Biostat, Chapel Hill, NC 27599 USA
[3] Univ N Carolina, Dept Stat & Operat Res, Stat, Chapel Hill, NC 27599 USA
基金
加拿大健康研究院; 美国国家卫生研究院; 美国国家科学基金会;
关键词
Heritability ratio; Imaging genetics; Multivariate regression; Projection regression; Sparse; Wild bootstrap; PRINCIPAL-COMPONENTS; BRAIN-DEVELOPMENT; MULTIVARIATE; CLASSIFICATION; FMRI; HERITABILITY; CONVERGENCE; RESPONSES; SELECTION; GENETICS;
D O I
10.1080/01621459.2014.892008
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The aim of this article is to develop a sparse projection regression modeling (SPReM) framework to perform multivariate regression modeling with a large number of responses and a multivariate covariate of interest. We propose two novel heritability ratios to simultaneously perform dimension reduction, response selection, estimation, and testing, while explicitly accounting for correlations among multivariate responses. Our SPReM is devised to specifically address the low statistical power issue of many standard statistical approaches, such as the Hotelling's T-2 test statistic or a mass univariate analysis, for high-dimensional data. We formulate the estimation problem of SPReM as a novel sparse unit rank projection (SURP) problem and propose a fast optimization algorithm for SURP. Furthermore, we extend SURP to the sparse multirank projection (SMURP) by adopting a sequential SURP approximation. Theoretically, we have systematically investigated the convergence properties of SURP and the convergence rate of SURP estimates. Our simulation results and real data analysis have shown that SPReM outperforms other state-of-the-art methods.
引用
收藏
页码:289 / 302
页数:14
相关论文
共 50 条
  • [41] Regularized estimation in sparse high-dimensional multivariate regression, with application to a DNA methylation study
    Zhang, Haixiang
    Zheng, Yinan
    Yoon, Grace
    Zhang, Zhou
    Gao, Tao
    Joyce, Brian
    Zhang, Wei
    Schwartz, Joel
    Vokonas, Pantel
    Colicino, Elena
    Baccarelli, Andrea
    Hou, Lifang
    Liu, Lei
    STATISTICAL APPLICATIONS IN GENETICS AND MOLECULAR BIOLOGY, 2017, 16 (03) : 159 - 171
  • [42] Sparse Bayesian variable selection in high-dimensional logistic regression models with correlated priors
    Ma, Zhuanzhuan
    Han, Zifei
    Ghosh, Souparno
    Wu, Liucang
    Wang, Min
    STATISTICAL ANALYSIS AND DATA MINING, 2024, 17 (01)
  • [43] Scalable High-Dimensional Multivariate Linear Regression for Feature-Distributed Data
    Huang, Shuo-Chieh
    Tsay, Ruey S.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25
  • [44] The Impact of Regularization on High-dimensional Logistic Regression
    Salehi, Fariborz
    Abbasi, Ehsan
    Hassibi, Babak
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [45] GENERALIZED REGRESSION ESTIMATORS WITH HIGH-DIMENSIONAL COVARIATES
    Ta, Tram
    Shao, Jun
    Li, Quefeng
    Wang, Lei
    STATISTICA SINICA, 2020, 30 (03) : 1135 - 1154
  • [46] SEMIPARAMETRIC QUANTILE REGRESSION WITH HIGH-DIMENSIONAL COVARIATES
    Zhu, Liping
    Huang, Mian
    Li, Runze
    STATISTICA SINICA, 2012, 22 (04) : 1379 - 1401
  • [47] Interpolating Predictors in High-Dimensional Factor Regression
    Bunea, Florentina
    Strimas-Mackey, Seth
    Wegkamp, Marten
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [48] Robust Ridge Regression for High-Dimensional Data
    Maronna, Ricardo A.
    TECHNOMETRICS, 2011, 53 (01) : 44 - 53
  • [49] Debiased inference for heterogeneous subpopulations in a high-dimensional logistic regression model
    Kim, Hyunjin
    Lee, Eun Ryung
    Park, Seyoung
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [50] A new accelerated proximal technique for regression with high-dimensional datasets
    Verma, Mridula
    Shukla, K. K.
    KNOWLEDGE AND INFORMATION SYSTEMS, 2017, 53 (02) : 423 - 438