MODEL-BASED SOFTWARE EFFORT ESTIMATION - A ROBUST COMPARISON OF 14 ALGORITHMS WIDELY USED IN THE DATA SCIENCE COMMUNITY

被引:8
作者
Phannachitta, Passakorn [1 ]
Matsumoto, Kenichi [2 ]
机构
[1] Chiang Mai Univ, Coll Arts Media & Technol, 239 Suthep, Chiang Mai 50200, Thailand
[2] Nara Inst Sci & Technol, Grad Sch Sci & Technol, 8916-5 Takayama, Ikoma, Nara 6300192, Japan
来源
INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL | 2019年 / 15卷 / 02期
关键词
Software effort estimation; Data science; Kaggle; Robust statistics; Empirical software engineering; COST ESTIMATION; PREDICTION; VALIDATION; REGRESSION; DIVERSITY;
D O I
10.24507/ijicic.15.02.569
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emergence of the data science discipline has facilitated the development of novel and advanced machine-learning algorithms for tackling tasks related to data analytics. For example, ensemble learning and deep learning have frequently achieved promising results in many recent data-science competitions, such as those hosted by Kaggle. However, these algorithms have not yet been thoroughly assessed on their performance when applied to software effort estimation. In this study, an assessment framework known as a stable-ranking-indication method is adopted to compare 14 machine-learning algorithms widely adopted in the data science communities. The comparisons were carried out over 13 industrial datasets, subject to six robust and independent performance metrics, and supported by the Brunner statistical test method. The results of this study proved to be stable because similar machine-learning algorithms achieved similar performance results; particularly, random forest and bagging performed the best among the compared algorithms. The results further offered evidence that demonstrated how to build an effective stacked ensemble. In other words, the optimal approach to maximizing the overall expected performance of the stacked ensemble can be derived through a balanced trade-off between maximizing the expected accuracy by selecting only the solo algorithms that are most likely to perform outstandingly on the dataset, and maximizing the level of diversity of the algorithms. Precisely, the stack combining bagging, random forests, analogy-based estimation, adaBoost, the gradient boosting machine, and ordinary least squares regression was shown to be the optimal stack for the software effort estimation datasets.
引用
收藏
页码:569 / 589
页数:21
相关论文
共 51 条
[1]   SOFTWARE FUNCTION, SOURCE LINES OF CODE, AND DEVELOPMENT EFFORT PREDICTION - A SOFTWARE SCIENCE VALIDATION [J].
ALBRECHT, AJ ;
GAFFNEY, JE .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 1983, 9 (06) :639-648
[2]  
[Anonymous], 2012, P INT C PRED MOD SOF
[3]  
[Anonymous], 2011, Modern Statistics for the Social and Behavioral Sciences: A Practical Introduction
[4]  
[Anonymous], 2017, IEEE T SOFTWARE ENG
[5]   A new perspective on data homogeneity in software cost estimation: a study in the embedded systems domain [J].
Bakir, Ayse ;
Turhan, Burak ;
Bener, Ayse B. .
SOFTWARE QUALITY JOURNAL, 2010, 18 (01) :57-80
[6]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[7]  
Brown G, 2005, J MACH LEARN RES, V6, P1621
[8]  
Brown G., 2005, Information Fusion, V6, P5, DOI 10.1016/j.inffus.2004.04.004
[9]  
Brown G, 2010, LECT NOTES COMPUT SC, V5997, P124, DOI 10.1007/978-3-642-12127-2_13
[10]   The multivariate nonparametric Behrens-Fisher problem [J].
Brunner, E ;
Munzel, U ;
Puri, ML .
JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2002, 108 (1-2) :37-53