An Adaptive Multiple Feature Subset Method for Feature Ranking and Selection

被引:4
作者
Chang, Fu [1 ]
Chen, Jen-Cheng [1 ]
机构
[1] Acad Sinica, Inst Informat Sci, Taipei, Taiwan
来源
INTERNATIONAL CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE (TAAI 2010) | 2010年
关键词
AMFES; CORR; curse of dimensionality; embedded method; feature ranking; feature selection; filter; RFE; wrapper; RANDOM SUBSPACE METHOD; BOUND ALGORITHM; CLASSIFICATION; INFORMATION; BRANCH;
D O I
10.1109/TAAI.2010.50
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a new feature evaluation method that forms the basis for feature ranking and selection. The method starts by generating a number of feature subsets in a random fashion and evaluates features based on the derived subsets. It then proceeds in a number of stages. In each stage, it inputs the features whose ranks in the previous stage were above the median rank and re-evaluates those features in the same fashion as it did in the first stage. When the number of features is high, the method has a computational advantage over recursive feature elimination (RFE), a state-of-art method that ranks features by identifying the least valuable feature in each stage. It also achieves better results than RFE in terms of classification accuracy and some other measures introduced in this paper, especially when the size of the training data is small or the number of irrelevant features is large.
引用
收藏
页码:255 / 262
页数:8
相关论文
共 51 条
[1]  
ALMUALLIM H, 1991, PROCEEDINGS : NINTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOLS 1 AND 2, P547
[2]   On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems [J].
Amaldi, E ;
Kann, V .
THEORETICAL COMPUTER SCIENCE, 1998, 209 (1-2) :237-260
[3]  
[Anonymous], 1989, P C ADV NEUR INF PRO
[4]  
Bi J., 2003, Journal of Machine Learning Research, V3, P1229, DOI 10.1162/153244303322753643
[5]   Selection of relevant features and examples in machine learning [J].
Blum, AL ;
Langley, P .
ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) :245-271
[6]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[7]  
Breiman L., 1984, Classification and regression trees, DOI DOI 10.1201/9781315139470
[8]  
Cardie C., 1993, P 10 INT C MACH LEAR, P25, DOI DOI 10.5555/3091529.3091533
[9]  
Chang Y.-W., 2008, Causation and Prediction Challenge, P53
[10]   Choosing multiple parameters for support vector machines [J].
Chapelle, O ;
Vapnik, V ;
Bousquet, O ;
Mukherjee, S .
MACHINE LEARNING, 2002, 46 (1-3) :131-159