BASSUM: A Bayesian semi-supervised method for classification feature selection

被引:31
作者
Cai, Ruichu [1 ]
Zhang, Zhenjie [2 ]
Hao, Zhifeng [1 ]
机构
[1] Guangdong Univ Technol, Fac Comp Sci, Guangzhou 510006, Guangdong, Peoples R China
[2] Natl Univ Singapore, Sch Comp, Singapore 117417, Singapore
关键词
Feature selection; Semi-supervised; Structured object; Markov blanket; Conditional independence test; GENE SELECTION;
D O I
10.1016/j.patcog.2010.10.023
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection is an important preprocessing step for building efficient, generalizable and interpretable classifiers on high dimensional data sets. Given the assumption on the sufficient labelled samples, the Markov Blanket provides a complete and sound solution to the selection of optimal features, by exploring the conditional independence relationships among the features. In real-world applications, unfortunately, it is usually easy to get unlabelled samples, but expensive to obtain the corresponding accurate labels on the samples. This leads to the potential waste of valuable classification information buried in unlabelled samples. In this paper, we propose a new BAyesian Semi-SUpervised Method, or BASSUM in short, to exploit the values of unlabelled samples on classification feature selection problem. Generally speaking, the inclusion of unlabelled samples helps the feature selection algorithm on (1) pinpointing more specific conditional independence tests involving fewer variable features and (2) improving the robustness of individual conditional independence tests with additional statistical information. Our experimental results show that BASSUM enhances the efficiency of traditional feature selection methods and overcomes the difficulties on redundant features in existing semi-supervised solutions. (C) 2010 Elsevier Ltd. All rights reserved.
引用
收藏
页码:811 / 820
页数:10
相关论文
共 20 条
[1]  
ALIFERIS CF, 2003, P 2003 AM MED INF AS, P21
[2]  
[Anonymous], P INT C MACH LEARN
[3]  
[Anonymous], P SIAM INT C DAT MIN
[4]  
[Anonymous], 1997, MACHINE LEARNING, MCGRAW-HILL SCIENCE/ENGINEERING/MATH
[5]   A BAYESIAN METHOD FOR THE INDUCTION OF PROBABILISTIC NETWORKS FROM DATA [J].
COOPER, GF ;
HERSKOVITS, E .
MACHINE LEARNING, 1992, 9 (04) :309-347
[6]   Structured machine learning: the next ten years [J].
Dietterich, Thomas G. ;
Domingos, Pedro ;
Getoor, Lise ;
Muggleton, Stephen ;
Tadepalli, Prasad .
MACHINE LEARNING, 2008, 73 (01) :3-23
[7]  
Dougherty, 1995, P 12 INT C MACH LEAR, V1995, P194, DOI DOI 10.1016/B978-1-55860-377-6.50032-3
[8]   Gene selection for cancer classification using support vector machines [J].
Guyon, I ;
Weston, J ;
Barnhill, S ;
Vapnik, V .
MACHINE LEARNING, 2002, 46 (1-3) :389-422
[9]  
HANDL J, 2006, P INT JOINT C NEUR N
[10]   Wrappers for feature subset selection [J].
Kohavi, R ;
John, GH .
ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) :273-324