On the consistency of Bayesian variable selection for high dimensional binary regression and classification

被引:7
作者
Jiang, Wenxin [1 ]
机构
[1] Northwestern Univ, Dept Stat, Evanston, IL 60208 USA
关键词
D O I
10.1162/neco.2006.18.11.2762
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Modern data mining and bioinformatics have presented an important playground for statistical learning techniques, where the number of input variables is possibly much larger than the sample size of the training data. In supervised learning, logistic regression or probit regression can be used to model a binary output and form perceptron classification rules based on Bayesian inference. We use a prior to select a limited number of candidate variables to enter the model, applying a popular method with selection indicators. We show that this approach can induce posterior estimates of the regression functions that are consistently estimating the truth, if the true regression model is sparse in the sense that the aggregated size of the regression coefficients are bounded. The estimated regression functions therefore can also produce consistent classifiers that are asymptotically optimal for predicting future binary outputs. These provide theoretical justifications for some recent empirical successes in microarray data analysis.
引用
收藏
页码:2762 / 2776
页数:15
相关论文
共 17 条
  • [1] BUHLMANN P, 2004, BOOSTING HIGH DIMENS
  • [2] Donoho D. L., 1993, Applied and Computational Harmonic Analysis, V1, P100, DOI 10.1006/acha.1993.1008
  • [3] On consistency of Bayesian inference with mixtures of logistic regression
    Ge, Y
    Jiang, WX
    [J]. NEURAL COMPUTATION, 2006, 18 (01) : 224 - 243
  • [4] Convergence rates of posterior distributions
    Ghosal, S
    Ghosh, JK
    Van der Vaart, AW
    [J]. ANNALS OF STATISTICS, 2000, 28 (02) : 500 - 531
  • [5] Asymptotic normality of posterior distributions in high-dimensional linear models
    Ghosal, S
    [J]. BERNOULLI, 1999, 5 (02) : 315 - 331
  • [6] Ghosal Subhashis, 1997, MATH METHODS STAT, V6, P332
  • [7] Persistence in high-dimensional linear predictor selection and the virtue of overparametrization
    Greenshtein, E
    Ritov, Y
    [J]. BERNOULLI, 2004, 10 (06) : 971 - 988
  • [8] Consistency of posterior distributions for neural networks
    Lee, HKH
    [J]. NEURAL NETWORKS, 2000, 13 (06) : 629 - 642
  • [9] Gene selection: a Bayesian variable selection approach
    Lee, KE
    Sha, NJ
    Dougherty, ER
    Vannucci, M
    Mallick, BK
    [J]. BIOINFORMATICS, 2003, 19 (01) : 90 - 97
  • [10] McCullagh P., 1989, GEN LINEAR MODELS, V2nd edn, DOI [DOI 10.1007/978-1-4899-3242-6, 10.1007/978-1-4899-3242-6, DOI 10.2307/2347392, 10.1201/9780203753736]