Sparse Discriminant Analysis

被引:387
作者
Clemmensen, Line [1 ]
Hastie, Trevor [2 ]
Witten, Daniela [3 ]
Ersboll, Bjarne [1 ]
机构
[1] Tech Univ Denmark, Dept Informat & Math Modelling, DK-2800 Lyngby, Denmark
[2] Stanford Univ, Dept Stat, Stanford, CA 94305 USA
[3] Univ Washington, Dept Biostat, Seattle, WA 98195 USA
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
Classification; Dimension reduction; Feature selection; Linear discriminant analysis; Mixture discriminant analysis; PARTIAL LEAST-SQUARES; REGRESSION; CLASSIFICATION; PREDICTION; DIAGNOSIS; SELECTION;
D O I
10.1198/TECH.2011.08118
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider the problem of performing interpretable classification in the high-dimensional setting, in which the number of features is very large and the number of observations is limited. This setting has been studied extensively in the chemometrics literature, and more recently has become commonplace in biological and medical applications. In this setting, a traditional approach involves performing feature selection before classification. We propose sparse discriminant analysis, a method for performing linear discriminant analysis with a sparseness criterion imposed such that classification and feature selection are performed simultaneously. Sparse discriminant analysis is based on the optimal scoring interpretation of linear discriminant analysis, and can be extended to perform sparse discrimination via mixtures of Gaussians if boundaries between classes are nonlinear or if subgroups are present within each class. Our proposal also provides low-dimensional views of the discriminative directions.
引用
收藏
页码:406 / 413
页数:8
相关论文
共 28 条
[1]   Partial least squares for discrimination [J].
Barker, M ;
Rayens, W .
JOURNAL OF CHEMOMETRICS, 2003, 17 (03) :166-173
[2]   Some theory for Fisher's linear discriminant function, 'naive Bayes', and some alternatives when there are many more variables than observations [J].
Bickel, PJ ;
Levina, E .
BERNOULLI, 2004, 10 (06) :989-1010
[3]   Sparse partial least squares regression for simultaneous dimension reduction and variable selection [J].
Chun, Hyonho ;
Keles, Suenduez .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2010, 72 :3-25
[4]   A method for comparison of growth media in objective identification of Penicillium based on multi-spectral imaging [J].
Clemmensen, Line H. ;
Hansen, Michael E. ;
Frisvad, Jens C. ;
Ersboll, Bjarne K. .
JOURNAL OF MICROBIOLOGICAL METHODS, 2007, 69 (02) :249-255
[5]  
CRAN, 2009, COMPR ARCH NETW
[6]  
Dudoit S., 2001, J. Am. Stat. Assoc, V96, P1151
[7]   Least angle regression - Rejoinder [J].
Efron, B ;
Hastie, T ;
Johnstone, I ;
Tibshirani, R .
ANNALS OF STATISTICS, 2004, 32 (02) :494-499
[8]   PATHWISE COORDINATE OPTIMIZATION [J].
Friedman, Jerome ;
Hastie, Trevor ;
Hoefling, Holger ;
Tibshirani, Robert .
ANNALS OF APPLIED STATISTICS, 2007, 1 (02) :302-332
[9]   REGULARIZED DISCRIMINANT-ANALYSIS [J].
FRIEDMAN, JH .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1989, 84 (405) :165-175
[10]   Interpretable Classifiers for fMRI Improve Prediction of Purchases [J].
Grosenick, Logan ;
Greer, Stephanie ;
Knutson, Brian .
IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2008, 16 (06) :539-548