Bias and stability of single variable classifiers for feature ranking and selection

被引:38
作者
Fakhraei, Shobeir [1 ,2 ]
Soltanian-Zadeh, Hamid [1 ,3 ]
Fotouhi, Farshad [4 ]
机构
[1] Henry Ford Hlth Syst, Dept Radiol, Med Image Anal Lab, Detroit, MI 48202 USA
[2] Univ Maryland, Dept Comp Sci, College Pk, MD 20740 USA
[3] Univ Tehran, Sch Elect & Comp Engn, Control & Intelligent Proc Ctr Excellence, Tehran 14395, Iran
[4] Wayne State Univ, Coll Engn, Detroit, MI 48202 USA
关键词
Feature ranking; Feature selection; Bias; Stability; Single variable classifier; Dimension reduction; Support Vector Machines; Naive Bayes; Multilayer Perceptron; K-Nearest Neighbors; Logistic Regression; Ada Boost; Random Forests; GENE SELECTION; ALGORITHMS;
D O I
10.1016/j.eswa.2014.05.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature rankings are often used for supervised dimension reduction especially when discriminating power of each feature is of interest, dimensionality of dataset is extremely high, or computational power is limited to perform more complicated methods. In practice, it is recommended to start dimension reduction via simple methods such as feature rankings before applying more complex approaches. Single variable classifier (SVC) ranking is a feature ranking based on the predictive performance of a classifier built using only a single feature. While benefiting from capabilities of classifiers, this ranking method is not as computationally intensive as wrappers. In this paper, we report the results of an extensive study on the bias and stability of such feature ranking method. We study whether the classifiers influence the SVC rankings or the discriminative power of features themselves has a dominant impact on the final rankings. We show the common intuition of using the same classifier for feature ranking and final classification does not always result in the best prediction performance. We then study if heterogeneous classifiers ensemble approaches provide more unbiased rankings and if they improve final classification performance. Furthermore, we calculate an empirical prediction performance loss for using the same classifier in SVC feature ranking and final classification from the optimal choices. (C) 2014 Elsevier Ltd. All rights reserved.
引用
收藏
页码:6945 / 6958
页数:14
相关论文
共 57 条
[1]   An Empirical Study on Wrapper-based Feature Ranking [J].
Altidor, Wilker ;
Khoshgoftaar, Taghi M. ;
Van Hulse, Jason .
ICTAI: 2009 21ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, 2009, :75-82
[2]  
[Anonymous], 2014, ADV KNOWLEDGE DISCOV, V527, P113
[3]  
[Anonymous], 2013, RECENT TRENDS APPL A, V7906, P332
[4]  
[Anonymous], 2006, Pattern recognition and machine learning
[5]   Empirical study of feature selection methods based on individual feature evaluation for classification problems [J].
Arauzo-Azofra, Antonio ;
Aznarte, Jose Luis ;
Benitez, Jose M. .
EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (07) :8170-8177
[6]   A feature selection technique for generation of classification committees and its application to categorization of laryngeal images [J].
Bacauskiene, M. ;
Verikas, A. ;
Gelzinis, A. ;
Valincius, D. .
PATTERN RECOGNITION, 2009, 42 (05) :645-654
[7]   Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets [J].
Bryll, R ;
Gutierrez-Osuna, R ;
Quek, F .
PATTERN RECOGNITION, 2003, 36 (06) :1291-1302
[8]   Combination of feature selection approaches with SVM in credit scoring [J].
Chen, Fei-Long ;
Li, Feng-Chia .
EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (07) :4902-4909
[9]  
Chen M., 2008, Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, P124, DOI DOI 10.1145/1401890.1401910
[10]   Combining multiple classifiers for wrapper feature selection [J].
Chrysostomou, Kyriacos ;
Chen, Sherry Y. ;
Liu, Xiaohui .
INTERNATIONAL JOURNAL OF DATA MINING MODELLING AND MANAGEMENT, 2008, 1 (01) :91-102