Seed Selection for Testing Deep Neural Networks

被引:7
作者
Zhi, Yuhan [1 ]
Xie, Xiaofei [2 ]
Shen, Chao [1 ]
Sun, Jun [2 ]
Zhang, Xiaoyu [1 ]
Guan, Xiaohong [1 ]
机构
[1] Xi An Jiao Tong Univ, 28 Xianning West Rd, Xian 710049, Peoples R China
[2] Singapore Management Univ, 81 Victoria St, Singapore 188065, Singapore
基金
中国国家自然科学基金; 新加坡国家研究基金会;
关键词
Deep learning testing; seed selection; coverage; robustness; TEST DATA GENERATION;
D O I
10.1145/3607190
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep learning (DL) has been applied in many applications. Meanwhile, the quality of DL systems is becoming a big concern. To evaluate the quality of DL systems, a number of DL testing techniques have been proposed. To generate test cases, a set of initial seed inputs are required. Existing testing techniques usually construct seed corpus by randomly selecting inputs from training or test dataset. Till now, there is no study on how initial seed inputs affect the performance of DL testing and how to construct an optimal one. To fill this gap, we conduct the first systematic study to evaluate the impact of seed selection strategies on DL testing. Specifically, considering three popular goals of DL testing (i.e., coverage, failure detection, and robustness), we develop five seed selection strategies, including three based on single-objective optimization (SOO) and two based on multi-objective optimization (MOO). We evaluate these strategies on seven testing tools. Our results demonstrate that the selection of initial seed inputs greatly affects the testing performance. SOO-based selection can construct the best seed corpus that can boost DL testing with respect to the specific testing goal. MOO-based selection strategies can construct seed corpus that achieve balanced improvement on multiple objectives.
引用
收藏
页数:33
相关论文
共 96 条
[71]   DeepConcolic: Testing and Debugging Deep Neural Networks [J].
Sun, Youcheng ;
Huang, Xiaowei ;
Kroening, Daniel ;
Sharp, James ;
Hill, Matthew ;
Ashmore, Rob .
2019 IEEE/ACM 41ST INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: COMPANION PROCEEDINGS (ICSE-COMPANION 2019), 2019, :111-114
[72]   Structural Test Coverage Criteria for Deep Neural Networks [J].
Sun, Youcheng ;
Huang, Xiaowei ;
Kroening, Daniel ;
Sharp, James ;
Hill, Matthew ;
Ashmore, Rob .
ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2019, 18 (05)
[73]   Concolic Testing for Deep Neural Networks [J].
Sun, Youcheng ;
Wu, Min ;
Ruan, Wenjie ;
Huang, Xiaowei ;
Kwiatkowska, Marta ;
Kroening, Daniel .
PROCEEDINGS OF THE 2018 33RD IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMTED SOFTWARE ENGINEERING (ASE' 18), 2018, :109-119
[74]  
Swiecki R., 2016, HONGGFUZZ
[75]   DeepTest: Automated Testing of Deep-Neural-Network-driven Autonomous Cars [J].
Tian, Yuchi ;
Pei, Kexin ;
Jana, Suman ;
Ray, Baishakhi .
PROCEEDINGS 2018 IEEE/ACM 40TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE), 2018, :303-314
[76]  
Tonella P., 2004, Software Engineering Notes, V29, P119, DOI 10.1145/1013886.1007528
[77]   DeepPath: Path-driven Testing Criteria for Deep Neural Networks [J].
Wang, Dong ;
Wang, Ziyuan ;
Fang, Chunrong ;
Chen, Yanshan ;
Chen, Zhenyu .
2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE TESTING (AITEST), 2019, :119-120
[78]   RobOT: Robustness-Oriented Testing for Deep Learning Systems [J].
Wang, Jingyi ;
Chen, Jialuo ;
Sun, Youcheng ;
Ma, Xingjun ;
Wang, Dongxia ;
Sun, Jun ;
Cheng, Peng .
2021 IEEE/ACM 43RD INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2021), 2021, :300-311
[79]   Skyfire: Data-Driven Seed Generation for Fuzzing [J].
Wang, Junjie ;
Chen, Bihuan ;
Wei, Lei ;
Liu, Yang .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :579-594
[80]   Superion: Grammar-Aware Greybox Fuzzing [J].
Wang, Junjie ;
Chen, Bihuan ;
Wei, Lei ;
Liu, Yang .
2019 IEEE/ACM 41ST INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2019), 2019, :724-735