Likelihood-Based Selection and Sharp Parameter Estimation

被引:193
作者
Shen, Xiaotong [1 ]
Pan, Wei [2 ]
Zhu, Yunzhang [1 ]
机构
[1] Univ Minnesota, Sch Stat, Minneapolis, MN 55455 USA
[2] Univ Minnesota, Div Biostat, Minneapolis, MN 55455 USA
基金
美国国家科学基金会;
关键词
Continuous but nonsmooth minimization; Coordinate descent; General likelihood; Graphical models; Nonconvex; (p; n)-Asymptotics; NONCONCAVE PENALIZED LIKELIHOOD; VARIABLE SELECTION; MODEL SELECTION; SPARSE; GRAPHS;
D O I
10.1080/01621459.2011.645783
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In high-dimensional data analysis, feature selection becomes one effective means for dimension reduction, which proceeds with parameter estimation. Concerning accuracy of selection and estimation, we study nonconvex constrained and regularized likelihoods in the presence of nuisance parameters. Theoretically, we show that constrained L-0 likelihood and its computational surrogate are optimal in that they achieve feature selection consistency and sharp parameter estimation, under one necessary condition required for any method to be selection consistent and to achieve sharp parameter estimation. It permits up to exponentially many candidate features. Computationally, we develop difference convex methods to implement the computational surrogate through prime and dual subproblems. These results establish a central role of L-0 constrained and regularized likelihoods in feature selection and parameter estimation involving selection. As applications of the general method and theory, we perform feature selection in linear regression and logistic regression, and estimate a precision matrix in Gaussian graphical models. In these situations, we gain a new theoretical insight and obtain favorable numerical results. Finally, we discuss an application to predict the metastasis status of breast cancer patients with their gene expression profiles. This article has online supplementary material.
引用
收藏
页码:223 / 232
页数:10
相关论文
共 29 条
[1]  
[Anonymous], AMERICAL MATHEMATICA
[2]  
Banerjee O, 2008, J MACH LEARN RES, V9, P485
[3]   SIMULTANEOUS ANALYSIS OF LASSO AND DANTZIG SELECTOR [J].
Bickel, Peter J. ;
Ritov, Ya'acov ;
Tsybakov, Alexandre B. .
ANNALS OF STATISTICS, 2009, 37 (04) :1705-1732
[4]  
Boyd S., 2004, CONVEX OPTIMIZATION, VFirst, DOI DOI 10.1017/CBO9780511804441
[5]   Extended Bayesian information criteria for model selection with large model spaces [J].
Chen, Jiahua ;
Chen, Zehua .
BIOMETRIKA, 2008, 95 (03) :759-771
[6]   NETWORK EXPLORATION VIA THE ADAPTIVE LASSO AND SCAD PENALTIES [J].
Fan, Jianqing ;
Feng, Yang ;
Wu, Yichao .
ANNALS OF APPLIED STATISTICS, 2009, 3 (02) :521-541
[7]   Variable selection via nonconcave penalized likelihood and its oracle properties [J].
Fan, JQ ;
Li, RZ .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2001, 96 (456) :1348-1360
[8]   Sparse inverse covariance estimation with the graphical lasso [J].
Friedman, Jerome ;
Hastie, Trevor ;
Tibshirani, Robert .
BIOSTATISTICS, 2008, 9 (03) :432-441
[9]   Recovering Sparse Signals With a Certain Family of Nonconvex Penalties and DC Programming [J].
Gasso, Gilles ;
Rakotomamonjy, Alain ;
Canu, Stephane .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2009, 57 (12) :4686-4698
[10]   CancerGenes: a gene selection resource for cancer genome projects [J].
Higgins, Maureen E. ;
Claremont, Martine ;
Major, John E. ;
Sander, Chris ;
Lash, Alex E. .
NUCLEIC ACIDS RESEARCH, 2007, 35 :D721-D726