Efficient robust doubly adaptive regularized regression with applications

被引:10
作者
Karunamuni, Rohana J. [1 ]
Kong, Linglong [1 ]
Tu, Wei [1 ]
机构
[1] Univ Alberta, Dept Math & Stat Sci, Edmonton, AB T6G 2G1, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Regularized regression; variable selection; efficiency; robustness; NONCONCAVE PENALIZED LIKELIHOOD; STATE FUNCTIONAL CONNECTIVITY; VARIABLE SELECTION; QUANTILE REGRESSION; CINGULATE CORTEX; DIVERGING NUMBER; ESTIMATORS; SHRINKAGE; BREAKDOWN; LASSO;
D O I
10.1177/0962280218757560
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
We consider the problem of estimation and variable selection for general linear regression models. Regularized regression procedures have been widely used for variable selection, but most existing methods perform poorly in the presence of outliers. We construct a new penalized procedure that simultaneously attains full efficiency and maximum robustness. Furthermore, the proposed procedure satisfies the oracle properties. The new procedure is designed to achieve sparse and robust solutions by imposing adaptive weights on both the decision loss and the penalty function. The proposed method of estimation and variable selection attains full efficiency when the model is correct and, at the same time, achieves maximum robustness when outliers are present. We examine the robustness properties using the finite-sample breakdown point and an influence function. We show that the proposed estimator attains the maximum breakdown point. Furthermore, there is no loss in efficiency when there are no outliers or the error distribution is normal. For practical implementation of the proposed method, we present a computational algorithm. We examine the finite-sample and robustness properties using Monte Carlo studies. Two datasets are also analyzed.
引用
收藏
页码:2210 / 2226
页数:17
相关论文
共 66 条
[1]   The ADHD-200 Consortium: a model to advance the translational potential of neuroimaging in clinical neuroscience [J].
Acuna, Carlos .
FRONTIERS IN SYSTEMS NEUROSCIENCE, 2012, 6
[2]   SPARSE LEAST TRIMMED SQUARES REGRESSION FOR ANALYZING HIGH-DIMENSIONAL LARGE DATA SETS [J].
Alfons, Andreas ;
Croux, Christophe ;
Gelper, Sarah .
ANNALS OF APPLIED STATISTICS, 2013, 7 (01) :226-248
[3]  
[Anonymous], FOUND TRENDS MACH LE
[4]  
[Anonymous], 2018, Robust Statistics: Theory and Methods (with R)
[5]   Efficient Implementations of the Generalized Lasso Dual Path Algorithm [J].
Arnold, Taylor B. ;
Tibshirani, Ryan J. .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2016, 25 (01) :1-27
[6]   Weighted LAD-LASSO method for robust parameter estimation and variable selection in regression [J].
Arslan, Olcay .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2012, 56 (06) :1952-1965
[7]   The Neuro Bureau ADHD-200 Preprocessed repository [J].
Bellec, Pierre ;
Chu, Carlton ;
Chouinard-Decorte, Francois ;
Benhajali, Yassine ;
Margulies, Daniel S. ;
Craddock, R. Cameron .
NEUROIMAGE, 2017, 144 :275-286
[8]  
Bickel P. J., 1998, EFFICIENT ADAPTIVE E
[9]   Penalized composite quasi-likelihood for ultrahigh dimensional variable selection [J].
Bradic, Jelena ;
Fan, Jianqing ;
Wang, Weiwei .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2011, 73 :325-349
[10]   Cingulate-precuneus interactions: A new locus of dysfunction in adult attention-deficit/hyperactivity disorder [J].
Castellanos, F. Xavier ;
Margulies, Daniel S. ;
Kelly, Clare ;
Uddin, Lucina Q. ;
Ghaffari, Manely ;
Kirsch, Andrew ;
Shaw, David ;
Shehzad, Zarrar ;
Di Martino, Adriana ;
Biswal, Bharat ;
Sonuga-Barke, Edmund J. S. ;
Rotrosen, John ;
Adler, Lenard A. ;
Milham, Michael P. .
BIOLOGICAL PSYCHIATRY, 2008, 63 (03) :332-337