Machine Learning-based Test Case Prioritization using Hyperparameter Optimization

被引:5
作者
Khan, Md Asif [1 ]
Azim, Akramul [1 ]
Liscano, Ramiro [1 ]
Smith, Kevin [2 ]
Tauseef, Qasim [2 ]
Seferi, Gkerta [2 ]
Chang, Yee-Kang [3 ]
机构
[1] Ontario Tech Univ, Oshawa, ON, Canada
[2] IBM United Kingdom Ltd, Portsmouth, Hants, England
[3] IBM Canada Ltd, Markham, ON, Canada
来源
PROCEEDINGS OF THE 2024 IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATION OF SOFTWARE TEST, AST 2024 | 2024年
基金
加拿大自然科学与工程研究理事会;
关键词
hyperparameter optimization; test case prioritization; machine learning; continuous integration;
D O I
10.1145/3644032.3644467
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Continuous integration pipelines execute extensive automated test suites to validate new software builds. In this fast-paced development environment, delivering timely testing results to developers is critical to ensuring software quality. Test case prioritization (TCP) emerges as a pivotal solution, enabling the prioritization of faultprone test cases for immediate attention. Recent advancements in machine learning have showcased promising results in TCP, offering the potential to revolutionize how we optimize testing workflows. Hyperparameter tuning plays a crucial role in enhancing the performance of ML models. However, there needs to be more work investigating the effects of hyperparameter tuning on TCP. Therefore, we explore how optimized hyperparameters influence the performance of various ML classifiers, focusing on the Average Percentage of Faults Detected (APFD) metric. Through empirical analysis of ten real-world, large-scale, diverse datasets, we conduct a grid search-based tuning with 885 hyperparameter combinations for four machine learning models. Our results provide model-specific insights and demonstrate an average 15% improvement in model performance with hyperparameter tuning compared to default settings. We further explain how hyperparameter tuning improves precision (max = 1), recall (max = 0.9633), F1-score (max = 0.9662), and influences APFD value (max = 0.9835), indicating a direct connection between tuning and prioritization performance. Hence, this study underscores the importance of hyperparameter tuning in optimizing failure prediction models and their direct impact on prioritization performance.
引用
收藏
页码:125 / 135
页数:11
相关论文
共 25 条
[1]   Simpler Hyperparameter Optimization for Software Analytics: Why, How, When? [J].
Agrawal, Amritanshu ;
Yang, Xueqi ;
Agrawal, Rishabh ;
Yedida, Rahul ;
Shen, Xipeng ;
Menzies, Tim .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2022, 48 (08) :2939-2954
[2]  
Aydemir E., 2018, P ENG 33 ACM IEEE IN, P294
[3]   A multi-objective artificial immune algorithm for parameter optimization in support vector machine [J].
Aydin, Ilhan ;
Karakose, Mehmet ;
Akin, Erhan .
APPLIED SOFT COMPUTING, 2011, 11 (01) :120-129
[4]   Learning-to-Rank vs Ranking-to-Learn: Strategies for Regression Testing in Continuous Integration [J].
Bertolino, Antonia ;
Guerriero, Antonio ;
Miranda, Breno ;
Pietrantuono, Roberto ;
Russo, Stefano .
2020 ACM/IEEE 42ND INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2020), 2020, :1-12
[5]   Hyperparameter optimization: Foundations, algorithms, best practices, and open challenges [J].
Bischl, Bernd ;
Binder, Martin ;
Lang, Michel ;
Pielok, Tobias ;
Richter, Jakob ;
Coors, Stefan ;
Thomas, Janek ;
Ullmann, Theresa ;
Becker, Marc ;
Boulesteix, Anne-Laure ;
Deng, Difan ;
Lindauer, Marius .
WILEY INTERDISCIPLINARY REVIEWS-DATA MINING AND KNOWLEDGE DISCOVERY, 2023, 13 (02)
[6]   Optimizing Test Prioritization via Test Distribution Analysis [J].
Chen, Junjie ;
Lou, Yiling ;
Zhang, Lingming ;
Zhou, Jianyi ;
Wang, Xiaoleng ;
Hao, Dan ;
Zhang, Lu .
ESEC/FSE'18: PROCEEDINGS OF THE 2018 26TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, 2018, :656-667
[7]   Test case prioritization: A family of empirical studies [J].
Elbaum, S ;
Malishevsky, AG ;
Rothermel, G .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2002, 28 (02) :159-182
[8]   Empirically Evaluating Readily Available Information for Regression Test Optimization in Continuous Integration [J].
Elsner, Daniel ;
Hauer, Florian ;
Pretschner, Alexander ;
Reimer, Silke .
ISSTA '21: PROCEEDINGS OF THE 30TH ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, 2021, :491-504
[9]  
Kocaguneli E., 2012, P 16 INT SOFTW PROD, P181
[10]  
Kuhn D, 2018, Arxiv, DOI arXiv:1806.10961