Revisiting Unsupervised Learning for Defect Prediction

被引:95
作者
Fu, Wei [1 ]
Menzies, Tim [1 ]
机构
[1] NC State, Com Sci, Raleigh, NC 27606 USA
来源
ESEC/FSE 2017: PROCEEDINGS OF THE 2017 11TH JOINT MEETING ON FOUNDATIONS OF SOFTWARE ENGINEERING | 2017年
基金
美国国家科学基金会;
关键词
Data analytics for software engineering; software repository mining; empirical studies; defect prediction; SOFTWARE CHANGES; FAULTS; HISTORY;
D O I
10.1145/3106237.3106257
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Collecting quality data from software projects can be time-consuming and expensive. Hence, some researchers explore "unsupervised" approaches to quality prediction that does not require labelled data. An alternate technique is to use "supervised" approaches that learn models from project data labelled with, say, "defective" or "not-defective". Most researchers use these supervised models since, it is argued, they can exploit more knowledge of the projects. At FSE'16, Yang et al. reported startling results where unsupervised defect predictors outperformed supervised predictors for efort-aware just-in-time defect prediction. If confirmed, these results would lead to a dramatic simplification of a seemingly complex task (data mining) that is widely explored in the software engineering literature. This paper repeats and refutes those results as follows. (1) There is much variability in the efficacy of the Yang et al. predictors so even with their approach, some supervised data is required to prune weaker predictors away. (2) Their findings were grouped across N projects. When we repeat their analysis on a project-by-project basis, supervised predictors are seen to work better. Even though this paper rejects the specific conclusions of Yang et al., we still endorse their general goal. In our our experiments, supervised predictors did not perform outstandingly better than unsupervised ones for efort-aware just-in-time defect prediction. Hence, they may indeed be some combination of unsupervised learners to achieve comparable performance to supervised ones. We therefore encourage others to work in this promising area.
引用
收藏
页码:72 / 83
页数:12
相关论文
共 56 条
[1]  
Agrawal A, 2017, CoRR abs/1705.03697
[2]  
Akiyama F., 1971, IFIP C, P353
[3]  
[Anonymous], P 38 INT C SOFTW ENG
[4]  
[Anonymous], 2007, 3 INT WORKSH PRED MO
[5]  
Arisholm E., 2006, ISESE 06 P 2006 ACMI, P8
[6]   CONTROLLING THE FALSE DISCOVERY RATE - A PRACTICAL AND POWERFUL APPROACH TO MULTIPLE TESTING [J].
BENJAMINI, Y ;
HOCHBERG, Y .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 1995, 57 (01) :289-300
[7]   A METRICS SUITE FOR OBJECT-ORIENTED DESIGN [J].
CHIDAMBER, SR ;
KEMERER, CF .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 1994, 20 (06) :476-493
[8]  
D'Ambros Marco, 2010, Proceedings of the 2010 7th IEEE Working Conference on Mining Software Repositories (MSR 2010), P31, DOI 10.1109/MSR.2010.5463279
[9]   Tuning for software analytics: Is it really necessary? [J].
Fu, Wei ;
Menzies, Tim ;
Shen, Xipeng .
INFORMATION AND SOFTWARE TECHNOLOGY, 2016, 76 :135-146
[10]  
Fukushima T., 2014, P 11 WORK C MIN SOFT, P172, DOI DOI 10.1145/2597073.2597075