An Empirical Comparison of Similarity Measures for Abstract Test Case Prioritization

被引:15
作者
Huang, Rubing [1 ]
Zhou, Yunan [1 ]
Zong, Weiwen [1 ]
Towey, Dave [2 ]
Chen, Jinfu [1 ]
机构
[1] Jiangsu Univ, Sch Comp Sci & Commun Engn, Zhenjiang 212013, Peoples R China
[2] Univ Nottingham Ningbo China, Sch Comp Sci, Ningbo 315100, Zhejiang, Peoples R China
来源
2017 IEEE 41ST ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE (COMPSAC), VOL 1 | 2017年
基金
中国国家自然科学基金;
关键词
Software testing; test case prioritization; abstract test case; similarity; COVERAGE;
D O I
10.1109/COMPSAC.2017.271
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Test case prioritization (TCP) attempts to order test cases such that those which are more important, according to some criterion or measurement, are executed earlier. TCP has been applied in many testing situations, including, for example, regression testing. An abstract test case (also called a model input) is an important type of test case, and has been widely used in practice, such as in configurable systems and software product lines. Similarity-based test case prioritization (STCP) has been proven to be cost-effective for abstract test cases (ATCs), but because there are many similarity measures which could be used to evaluate ATCs and to support STCP, we face the following question: How can we choose the similarity measure(s) for prioritizing ATCs that will deliver the most effective results? To address this, we studied fourteen measures and two popular STCP algorithms - local STCP (LSTCP), and global STCP (GSTCP). We also conducted an empirical study of five real-world programs, and investigated the efficacy of each similarity measure, according to the interaction coverage rate and fault detection rate. The results of these studies show that GSTCP outperforms LSTCP - in 61% to 84% of the cases, in terms of interaction coverage rates; and in 76% to 78% of the cases with respect to fault detection rates. Our studies also show that Overlap, the simplest similarity measure examined in this study, could obtain the overall best performance for LSTCP; and that Goodall3 has the best performance for GSTCP.
引用
收藏
页码:3 / 12
页数:10
相关论文
共 23 条
[1]   Similarity-Based Prioritization in Software Product-Line Testing [J].
Al-Hajjaji, Mustafa ;
Thuem, Thomas ;
Meinicke, Jens ;
Lochau, Malte ;
Saake, Gunter .
18TH INTERNATIONAL SOFTWARE PRODUCT LINE CONFERENCE (SPLC 2014), VOL 1, 2014, :197-206
[2]   Using mutation analysis for assessing and comparing testing coverage criteria [J].
Andrews, James H. ;
Briand, Lionel C. ;
Labiche, Yvan ;
Namin, Akbar Siami .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2006, 32 (08) :608-624
[3]  
Boriah S., 2008, P 8 SIAM INT C DAT M, P243, DOI DOI 10.1137/1.9781611972788.22
[4]   Prioritized interaction testing for pair-wise coverage with seeding and constraints [J].
Bryce, Renee C. ;
Colbourn, Charles J. .
INFORMATION AND SOFTWARE TECHNOLOGY, 2006, 48 (10) :960-970
[5]   Supporting controlled experimentation with testing techniques: An infrastructure and its potential impact [J].
Do, HS ;
Elbaum, S ;
Rothermel, G .
EMPIRICAL SOFTWARE ENGINEERING, 2005, 10 (04) :405-435
[6]  
Do H, 2006, IEEE T SOFTWARE ENG, V32, P733, DOI 10.1109/TSE.2006.92
[7]   Categorical data clustering: What similarity measure to recommend? [J].
dos Santos, Tiago R. L. ;
Zarate, Luis E. .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (03) :1247-1260
[8]   An evaluation of combination strategies for test case selection [J].
Grindall, Mats ;
Lindstroem, Birgitta ;
Offutt, Jeff ;
Andler, Sten F. .
EMPIRICAL SOFTWARE ENGINEERING, 2006, 11 (04) :583-611
[9]   Achieving Scalable Model-Based Testing Through Test Case Diversity [J].
Hemmati, Hadi ;
Arcuri, Andrea ;
Briand, Lionel .
ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2013, 22 (01)
[10]   Comparing White-box and Black-box Test Prioritization [J].
Henard, Christopher ;
Papadakis, Mike ;
Harman, Mark ;
Jia, Yue ;
Le Traon, Yves .
2016 IEEE/ACM 38TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE), 2016, :523-534