Manifesting Bugs in Machine Learning Code: An Explorative Study with Mutation Testing

被引:24
作者
Cheng, Dawei [1 ]
Cao, Chun [1 ]
Xu, Chang [1 ]
Ma, Xiaoxing [1 ]
机构
[1] Nanjing Univ, Inst Comp Software, State Key Lab Novel Software Technol, Nanjing, Peoples R China
来源
2018 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE QUALITY, RELIABILITY AND SECURITY (QRS 2018) | 2018年
基金
国家重点研发计划;
关键词
machine learning programs; mutation testing; explorative study;
D O I
10.1109/QRS.2018.00044
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Nowadays statistical machine learning is widely adopted in various domains such as data mining, image recognition and automated driving. However, software quality assurance for machine learning is still in its infancy. While recent efforts have been put into improving the quality of training data and trained models, this paper focuses on code-level bugs in the implementations of machine learning algorithms. In this explorative study we simulated program bugs by mutating Weka implementations of several classification algorithms. We observed that 8%-40% of the logically non-equivalent executable mutants were statistically indistinguishable from their golden versions. Moreover, other 15%-36% of the mutants were stubborn, as they performed not significantly worse than a reference classifier on at least one natural data set. We also experimented with several approaches to killing those stubborn mutants. Preliminary results indicate that bugs in machine learning code may have negative impacts on statistical properties such as robustness and learning curves, but they could be very difficult to detect, due to the lack of effective oracles.
引用
收藏
页码:313 / 324
页数:12
相关论文
共 49 条
  • [31] Evaluation of Mutation Testing in a Nuclear Industry Case Study
    Delgado-Perez, Pedro
    Habli, Ibrahim
    Gregory, Steve
    Alexander, Rob
    Clark, John
    Medina-Bulo, Inmaculada
    IEEE TRANSACTIONS ON RELIABILITY, 2018, 67 (04) : 1406 - 1419
  • [32] To Kill a Mutant: An Empirical Study of Mutation Testing Kills
    Du, Hang
    Palepu, Vijay Krishna
    Jones, James A.
    PROCEEDINGS OF THE 32ND ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, ISSTA 2023, 2023, : 715 - 726
  • [33] Cost Measures Matter for Mutation Testing Study Validity
    Guizzo, Giovani
    Sarro, Federica
    Harman, Mark
    PROCEEDINGS OF THE 28TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING (ESEC/FSE '20), 2020, : 1127 - 1139
  • [34] Deep Learning for Hyperspectral Image Classification: A Critical Evaluation via Mutation Testing
    Chen, Zhifei
    Hao, Yang
    Liu, Qichao
    Liu, Yuyong
    Zhu, Mingyang
    Xiao, Liang
    REMOTE SENSING, 2024, 16 (24)
  • [35] Mutation Testing in Continuous Integration: An Exploratory Industrial Case Study
    Orgard, Jonathan
    Gay, Gregory
    de Oliveira Neto, Francisco Gomes
    Viggedal, Kim
    2023 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE TESTING, VERIFICATION AND VALIDATION WORKSHOPS, ICSTW, 2023, : 324 - 333
  • [36] An Extensive Study on Cross-project Predictive Mutation Testing
    Mao, Dongyu
    Chen, Lingchao
    Zhang, Lingming
    2019 IEEE 12TH CONFERENCE ON SOFTWARE TESTING, VALIDATION AND VERIFICATION (ICST 2019), 2019, : 160 - 171
  • [37] Test-driven development with mutation testing - an experimental study
    Roman, Adam
    Mnich, Michal
    SOFTWARE QUALITY JOURNAL, 2021, 29 (01) : 1 - 38
  • [38] Test-driven development with mutation testing – an experimental study
    Adam Roman
    Michal Mnich
    Software Quality Journal, 2021, 29 : 1 - 38
  • [39] Application of Mutation testing in SafetyCritical Embedded Systems: A Case Study
    Serban, Andrada A.
    Micskei, Zoltan
    ACTA POLYTECHNICA HUNGARICA, 2024, 21 (08)
  • [40] An Empirical Study on Mutation Testing of WS-BPEL Programs
    Sun, Chang-ai
    Pan, Lin
    Wang, Qiaoling
    Liu, Huai
    Zhang, Xiangyu
    COMPUTER JOURNAL, 2017, 60 (01) : 143 - 158