Robust adaptive LASSO in high-dimensional logistic regression

被引:1
作者
Basu, Ayanendranath [1 ]
Ghosh, Abhik [1 ]
Jaenada, Maria [2 ]
Pardo, Leandro [2 ]
机构
[1] Indian Stat Inst, Interdisciplinary Stat Res Unit, 203 BT Rd, Kolkata 700108, India
[2] Univ Complutense Madrid, Stat & OR, Plaza Ciencias 3, Madrid 28040, Spain
关键词
Density power divergence; High-dimensional data; Logistic regression; Oracle properties; Variable selection; VARIABLE SELECTION; GENE SELECTION; SPARSE REGRESSION; CLASSIFICATION; CANCER; MICROARRAYS; LIKELIHOOD; ALGORITHM; MODELS;
D O I
10.1007/s10260-024-00760-2
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Penalized logistic regression is extremely useful for binary classification with large number of covariates (higher than the sample size), having several real life applications, including genomic disease classification. However, the existing methods based on the likelihood loss function are sensitive to data contamination and other noise and, hence, robust methods are needed for stable and more accurate inference. In this paper, we propose a family of robust estimators for sparse logistic models utilizing the popular density power divergence based loss function and the general adaptively weighted LASSO penalties. We study the local robustness of the proposed estimators through its influence function and also derive its oracle properties and asymptotic distribution. With extensive empirical illustrations, we demonstrate the significantly improved performance of our proposed estimators over the existing ones with particular gain in robustness. Our proposal is finally applied to analyse four different real datasets for cancer classification, obtaining robust and accurate models, that simultaneously performs gene selection and patient classification.
引用
收藏
页码:1217 / 1249
页数:33
相关论文
共 53 条
[41]   A RIDGE LOGISTIC ESTIMATOR [J].
SCHAEFER, RL ;
ROI, LD ;
WOLFE, RA .
COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 1984, 13 (01) :99-113
[42]   A simple and efficient algorithm for gene selection using sparse logistic regression [J].
Shevade, SK ;
Keerthi, SS .
BIOINFORMATICS, 2003, 19 (17) :2246-2253
[43]   Penalized logistic regression for high-dimensional DNA methylation data with case-control studies [J].
Sun, Hokeun ;
Wang, Shuang .
BIOINFORMATICS, 2012, 28 (10) :1368-1375
[44]   Elastic Net Regularization Paths for All Generalized Linear Models [J].
Tay, J. Kenneth ;
Narasimhan, Balasubramanian ;
Hastie, Trevor .
JOURNAL OF STATISTICAL SOFTWARE, 2023, 106 (01) :1-31
[46]   Gene expression profiling predicts clinical outcome of breast cancer [J].
van't Veer, LJ ;
Dai, HY ;
van de Vijver, MJ ;
He, YDD ;
Hart, AAM ;
Mao, M ;
Peterse, HL ;
van der Kooy, K ;
Marton, MJ ;
Witteveen, AT ;
Schreiber, GJ ;
Kerkhoven, RM ;
Roberts, C ;
Linsley, PS ;
Bernards, R ;
Friend, SH .
NATURE, 2002, 415 (6871) :530-536
[47]   Robust Variable Selection With Exponential Squared Loss [J].
Wang, Xueqin ;
Jiang, Yunlu ;
Huang, Mian ;
Zhang, Heping .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2013, 108 (502) :632-643
[48]   Choosing a robustness tuning parameter [J].
Warwick, J ;
Jones, MC .
JOURNAL OF STATISTICAL COMPUTATION AND SIMULATION, 2005, 75 (07) :581-588
[49]   Predicting the clinical status of human breast cancer by using gene expression profiles [J].
West, M ;
Blanchette, C ;
Dressman, H ;
Huang, E ;
Ishida, S ;
Spang, R ;
Zuzan, H ;
Olson, JA ;
Marks, JR ;
Nevins, JR .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2001, 98 (20) :11462-11467
[50]   A sparse logistic regression framework by difference of convex functions programming [J].
Yang, Liming ;
Qian, Yannan .
APPLIED INTELLIGENCE, 2016, 45 (02) :241-254