On the Stability of Feature Selection Methods in Software Quality Prediction: An Empirical Investigation

被引:7
作者
Wang, Huanjing [1 ]
Khoshgoftaar, Taghi M. [2 ]
Seliya, Naeem [3 ]
机构
[1] Western Kentucky Univ, Dept Comp Sci, Bowling Green, KY 42101 USA
[2] Florida Atlantic Univ, Dept Comp & Elect Engn & Comp Sci, Boca Raton, FL 33431 USA
[3] Hlth Safety Technol LLC, Coosada, AL USA
基金
美国国家科学基金会;
关键词
Software metrics; feature selection; fixed-overlap partitions; stability; ALGORITHMS;
D O I
10.1142/S0218194015400288
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Software quality modeling is the process of using software metrics from previous iterations of development to locate potentially faulty modules in current under-development code. This has become an important part of the software development process, allowing practitioners to focus development efforts where they are most needed. One difficulty encountered in software quality modeling is the problem of high dimensionality, where the number of available software metrics is too large for a classifier to work well. In this case, many of the metrics may be redundant or irrelevant to defect prediction results, thereby selecting a subset of software metrics that are the best predictors becomes important. This process is called feature (metric) selection. There are three major forms of feature selection: filter-based feature rankers, which uses statistical measures to assign a score to each feature and present the user with a ranked list; filter-based feature subset evaluation, which uses statistical measures on feature subsets to find the best feature subset; and wrapper-based subset selection, which builds classification models using different subsets to find the one which maximizes performance. Software practitioners are interested in which feature selection methods are best at providing the most stable feature subset in the face of changes to the data (here, the addition or removal of instances). In this study we select feature subsets using fifteen feature selection methods and then use our newly proposed Average Pairwise Tanimoto Index (APTI) to evaluate the stability of the feature selection methods. We evaluate the stability of feature selection methods on a pair of subsamples generated by our fixed-overlap partitions algorithm. Four different levels of overlap are considered in this study. 13 software metric datasets from two real-world software projects are used in this study. Results demonstrate that ReliefF (RF) is the most stable feature selection method and wrapper based feature subset selection shows least stability. In addition, as the overlap of partitions increased, the stability of the feature selection strategies increased.
引用
收藏
页码:1467 / 1490
页数:24
相关论文
共 31 条
[1]  
Alelyani S., 2011, IEEE INT C HIGH PERF, P701
[2]  
[Anonymous], P 25 INT C SOFTW ENG
[3]  
[Anonymous], MACHINE LEARNING
[4]  
Berenson M.L., 1983, INTERMEDIATE STAT ME, DOI DOI 10.2307/2288297
[5]  
Dunne K., 2015, TCDCD200228, P1
[6]  
Feldt R, 2010, 22ND INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING & KNOWLEDGE ENGINEERING (SEKE 2010), P374
[7]   MSR 2007 4th International Workshop on Mining Software Repositories [J].
Gall, Harald ;
Lanza, Michele ;
Zimmermann, Thomas .
29TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: ICSE 2007 COMPANION VOLUME, PROCEEDINGS, 2007, :107-+
[8]  
Guyon Isabelle, 2003, Journal of Machine Learning, V3, P1157
[9]  
Hall M., 1997, THESIS U WAIKATO HAM
[10]   The Influence of Feature Selection Methods on Accuracy, Stability and Interpretability of Molecular Signatures [J].
Haury, Anne-Claire ;
Gestraud, Pierre ;
Vert, Jean-Philippe .
PLOS ONE, 2011, 6 (12)