Evaluating noise elimination techniques for software quality estimation

被引:2
作者
Khoshgoftaar, Taghi M. [1 ]
Rebours, Pierre [1 ]
机构
[1] Florida Atlantic Univ, Dept Comp Sci & Engn, Empirical Software Engn Lab, Boca Raton, FL 33431 USA
关键词
D O I
10.3233/IDA-2005-9506
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The poor quality of a training dataset can have untoward consequences in software quality estimation problems. The presence of noise in software measurement data may hinder the prediction accuracy of a given learner. A filter improves the quality of training datasets by removing data that is likely noise. We evaluate the Ensemble Filter against the Partitioning Filter and the Classification Filter. These filtering techniques combine the predictions of base classifiers in such a way that an instance is identified as noisy if it is misclassified by a given number of these learners. The Partitioning Filter first splits the training dataset into subsets, and different base learners are induced on each subset. Two different implementations of the Partitioning Filter are presented: the Multiple-Partitioning Filter and the Iterative-Partitioning Filter. In contrast, the Ensemble Filter uses base classifiers induced on the entire training dataset. The filtering level and/or the number of iterations modify the filtering conservativeness: a conservative filter is less likely to remove good data at the expense of retaining noisy instances. A unique measure for comparing the relative efficiencies of two filters is also presented. Empirical studies on a high assurance software project evaluate the relative performances of the Ensemble Filter, Multiple-Partitioning Filter, Iterative-Partitioning Filter, and Classification Filter. Our study demonstrates that with a conservative filtering approach, using several different base learners can improve the efficiency of the filtering schemes.
引用
收藏
页码:487 / 508
页数:22
相关论文
共 50 条
[41]   Survey of Software Development Effort Estimation Techniques [J].
Saeed, Ayesha ;
Butt, Wasi Haider ;
Kazmi, Farwa ;
Arif, Madeha .
PROCEEDINGS OF 2018 7TH INTERNATIONAL CONFERENCE ON SOFTWARE AND COMPUTER APPLICATIONS (ICSCA 2018), 2018, :82-86
[42]   SOFTWARE TECHNIQUES FOR EVALUATING KARTRAK SYSTEM PERFORMANCE. [J].
Anon .
Railway System Controls, 1975, 6 (02) :14-15
[43]   A Methodological Framework for Evaluating Software Testing Techniques and Tools [J].
Vos, Tanja E. J. ;
Marin, Beatriz ;
Jose Escalona, Maria ;
Marchetto, Alessandro .
2012 12TH INTERNATIONAL CONFERENCE ON QUALITY SOFTWARE (QSIC), 2012, :230-239
[44]   A systemic Quality Model for evaluating software products [J].
Ortega, M ;
Pérez, MA ;
Rojas, T .
6TH WORLD MULTICONFERENCE ON SYSTEMICS, CYBERNETICS AND INFORMATICS, VOL I, PROCEEDINGS: INFORMATION SYSTEMS DEVELOPMENT I, 2002, :371-376
[45]   A Software Processing Chain for Evaluating Thesaurus Quality [J].
Lacasta, Javier ;
Falquet, Gilles ;
Nogueras-Iso, Javier ;
Zarazaga-Soria, Javier .
SEMANTIC KEYWORD-BASED SEARCH ON STRUCTURED DATA SOURCES, IKC 2016, 2017, 10151 :91-99
[46]   Evaluating Estimation Techniques of Transportation Price Elasticity [J].
Han, Chien-Pai ;
Li, Jianling .
TRANSPORTATION RESEARCH RECORD, 2009, (2115) :94-101
[47]   An exercise in evaluating significance of software quality criteria [J].
Szejko, S .
ITICSE '99: PROCEEDINGS OF THE 4TH ANNUAL SIGCSE/SIGCUE CONFERENCE ON INNOVATION AND TECHNOLOGY IN COMPUTER SCIENCE EDUCATION, 1999, 31 (03) :199-199
[48]   Evaluating software quality mechanisms - A snapshot picture [J].
Hall, T .
SOFTWARE QUALITY MANAGEMENT IV - IMPROVING QUALITY, 1996, :251-260
[49]   Evaluating the Quality of the Documentation of Open Source Software [J].
Aversano, Lerina ;
Guardabascio, Daniela ;
Tortorella, Maria .
ENASE: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON EVALUATION OF NOVEL APPROACHES TO SOFTWARE ENGINEERING, 2017, :308-313
[50]   Hierarchical Model for Evaluating Software Design Quality [J].
Martenka, Pawel ;
Walter, Bartosz .
E-INFORMATICA SOFTWARE ENGINEERING JOURNAL, 2010, 4 (01) :21-30