On the dataset shift problem in software engineering prediction models

被引:7
作者
Burak Turhan
机构
[1] University of Oulu,Department of Information Processing Science
来源
Empirical Software Engineering | 2012年 / 17卷
关键词
Dataset shift; Prediction models; Effort estimation; Defect prediction;
D O I
暂无
中图分类号
学科分类号
摘要
A core assumption of any prediction model is that test data distribution does not differ from training data distribution. Prediction models used in software engineering are no exception. In reality, this assumption can be violated in many ways resulting in inconsistent and non-transferrable observations across different cases. The goal of this paper is to explain the phenomena of conclusion instability through the dataset shift concept from software effort and fault prediction perspective. Different types of dataset shift are explained with examples from software engineering, and techniques for addressing associated problems are discussed. While dataset shifts in the form of sample selection bias and imbalanced data are well-known in software engineering research, understanding other types is relevant for possible interpretations of the non-transferable results across different sites and studies. Software engineering community should be aware of and account for the dataset shift related issues when evaluating the validity of research outcomes.
引用
收藏
页码:62 / 74
页数:12
相关论文
共 53 条
[1]  
Bakır A(2010)A new perspective on data homogeneity in software cost estimation: a study in the embedded systems domain Softw Qual J 18 57-80
[2]  
Turhan B(2009)Discriminative learning under covariate shift J Mach Learn Res 10 2137-2155
[3]  
Bener A(2002)Empirical studies of quality models in object-oriented systems Adv Comput 56 97-166
[4]  
Bickel S(2002)Assessing the applicability of fault-proneness models across object-oriented software projects IEEE Trans Softw Eng 28 706-720
[5]  
Brückner M(2009)Anomaly detection: a survey ACM Comput Surv 41 15:1-15:58
[6]  
Scheffer T(2009)Conceptual association of functional size measurement methods IEEE Softw 26 71-78
[7]  
Briand L(2006)Cost curves: an improved method for visualizing classifier performance Mach Learn 65 95-130
[8]  
Wust J(2006)Classifier technology and the illusion of progress Stat Sci 21 1-15
[9]  
Briand LC(2008)Techniques for evaluating fault prediction models Empir Soft Eng 13 561-595
[10]  
Melo WL(2008)Analogy-X: providing statistical inference to analogy-based software cost estimation IEEE Trans Softw Eng 34 471-484