Testing of Deep Reinforcement Learning Agents with Surrogate Models

被引:8
作者
Biagiola, Matteo [1 ]
Tonella, Paolo [1 ]
机构
[1] Univ Svizzera Italiana, 6900 Via Buffi 13, Lugano, Switzerland
基金
欧盟地平线“2020”;
关键词
Software testing; reinforcement learning;
D O I
10.1145/3631970
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Deep Reinforcement Learning (DRL) has received a lot of attention from the research community in recent years. As the technology moves away from game playing to practical contexts, such as autonomous vehicles and robotics, it is crucial to evaluate the quality of DRL agents. In this article, we propose a search-based approach to test such agents. Our approach, implemented in a tool called Indago, trains a classifier on failure and non-failure environment (i.e., pass) configurations resulting from the DRL training process. The classifier is used at testing time as a surrogate model for the DRL agent execution in the environment, predicting the extent to which a given environment configuration induces a failure of the DRL agent under test. The failure prediction acts as a fitness function, guiding the generation towards failure environment configurations, while saving computation time by deferring the execution of the DRL agent in the environment to those configurations that are more likely to expose failures. Experimental results show that our search-based approach finds 50% more failures of the DRL agent than state-of-the-art techniques. Moreover, such failures are, on average, 78% more diverse; similarly, the behaviors of the DRL agent induced by failure configurations are 74% more diverse.
引用
收藏
页数:33
相关论文
共 50 条
[21]   Learning Sparse Evidence-Driven Interpretation to Understand Deep Reinforcement Learning Agents [J].
Dao, Giang ;
Huff, Wesley Houston ;
Lee, Minwoo .
2021 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2021), 2021,
[22]   The role of Reinforcement Learning in software testing [J].
Abo-eleneen, Amr ;
Palliyali, Ahammed ;
Catal, Cagatay .
INFORMATION AND SOFTWARE TECHNOLOGY, 2023, 164
[23]   Testing for Fault Diversity in Reinforcement Learning [J].
Mazouni, Quentin ;
Spieker, Helge ;
Gotlieb, Arnaud ;
Acher, Mathieu .
PROCEEDINGS OF THE 2024 IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATION OF SOFTWARE TEST, AST 2024, 2024, :136-146
[24]   Natural Walking With Musculoskeletal Models Using Deep Reinforcement Learning [J].
Weng, Jiacheng ;
Hashemi, Ehsan ;
Arami, Arash .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) :4156-4162
[25]   RLBoost: Boosting supervised models using deep reinforcement learning [J].
Batanero, Eloy Anguiano ;
Pascual, angela Fernandez ;
Jimenez, alvaro Barbero .
NEUROCOMPUTING, 2025, 618
[26]   A Hybrid Multi-Task Learning Approach for Optimizing Deep Reinforcement Learning Agents [J].
Varghese, Nelson Vithayathil ;
Mahmoud, Qusay H. .
IEEE ACCESS, 2021, 9 :44681-44703
[27]   Evading Machine Learning Botnet Detection Models via Deep Reinforcement Learning [J].
Wu, Di ;
Fang, Binxing ;
Wang, Junnan ;
Liu, Qixu ;
Cui, Xiang .
ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
[28]   Learning to Calibrate Battery Models in Real-Time with Deep Reinforcement Learning [J].
Unagar, Ajaykumar ;
Tian, Yuan ;
Chao, Manuel Arias ;
Fink, Olga .
ENERGIES, 2021, 14 (05)
[29]   Learning to Teach Reinforcement Learning Agents [J].
Fachantidis, Anestis ;
Taylor, Matthew ;
Vlahavas, Ioannis .
MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2019, 1 (01) :21-42
[30]   From Reinforcement Learning to Deep Reinforcement Learning: An Overview [J].
Agostinelli, Forest ;
Hocquet, Guillaume ;
Singh, Sameer ;
Baldi, Pierre .
BRAVERMAN READINGS IN MACHINE LEARNING: KEY IDEAS FROM INCEPTION TO CURRENT STATE, 2018, 11100 :298-328