Explainable reinforcement learning for powertrain control engineering

被引:1
作者
Laflamme, C. [1 ]
Doppler, J. [2 ]
Palvolgyi, B. [3 ]
Dominka, S. [2 ]
Viharos, Zs. J. [3 ,4 ]
Haeussler, S. [5 ]
机构
[1] Fraunhofer Austria Res GmbH, Vienna, Austria
[2] Robert Bosch AG, Bosch Engn, Vienna, Austria
[3] HUN REN Inst Comp Sci & Control SZTAK, Ctr Excellence Hungarian Acad Sci MTA, Budapest, Hungary
[4] John von Neumann Univ, Fac Econ & Business, Kecskemet, Hungary
[5] Univ Innsbruck, Dept Informat Syst Prod & Logist Management, Innsbruck, Austria
基金
欧盟地平线“2020”;
关键词
Reinforcement learning; Explainable artificial intelligence; Powertrain control; HYBRID ELECTRIC VEHICLE; ENERGY MANAGEMENT; RULES;
D O I
10.1016/j.engappai.2025.110135
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper we demonstrate a practical post-hoc approach for explainable reinforcement learning (RL) in vehicle powertrain control. The goal is to exploit the advantages of RL yet obtain a solution that is feasible to implement in safety-critical control engineering problems. This means finding a solution that balances optimal product design with the required engineering effort, while maintaining the transparency necessary for safety- critical applications. Our method is based on initially training a neural network based RL policy and converting it into a look-up table, using a decision tree (DT) as an intermediary. The DT is limited to a certain depth, resulting in a look-up table of manageable size that can be directly tested, implemented and evaluated by control engineers. In order to evaluate this approach, a set of RL expert policies were used to train DTs with increasing depth, showing the regions where the DT solution can outperform benchmarks while still remaining small enough to translate to a manageable look-up table. Our approach involves standard Python libraries, lowering the barrier for implementation. This approach is not just relevant to powertrain control, but offers a practical approach for all regulated domains which could benefit from application of RL.
引用
收藏
页数:12
相关论文
共 79 条
[1]  
Ahmad MA, 2018, ACM-BCB'18: PROCEEDINGS OF THE 2018 ACM INTERNATIONAL CONFERENCE ON BIOINFORMATICS, COMPUTATIONAL BIOLOGY, AND HEALTH INFORMATICS, P559, DOI [10.1145/3233547.3233667, 10.1109/ICHI.2018.00095]
[2]   Survey and critique of techniques for extracting rules from trained artificial neural networks [J].
Andrews, R ;
Diederich, J ;
Tickle, AB .
KNOWLEDGE-BASED SYSTEMS, 1995, 8 (06) :373-389
[3]   Survey of Deep Reinforcement Learning for Motion Planning of Autonomous Vehicles [J].
Aradi, Szilard .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (02) :740-759
[4]   Modeling dispositional and initial learned trust in automated vehicles with predictability and explainability [J].
Ayoub, Jackie ;
Yang, X. Jessie ;
Zhou, Feng .
TRANSPORTATION RESEARCH PART F-TRAFFIC PSYCHOLOGY AND BEHAVIOUR, 2021, 77 :102-116
[5]   Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI [J].
Barredo Arrieta, Alejandro ;
Diaz-Rodriguez, Natalia ;
Del Ser, Javier ;
Bennetot, Adrien ;
Tabik, Siham ;
Barbado, Alberto ;
Garcia, Salvador ;
Gil-Lopez, Sergio ;
Molina, Daniel ;
Benjamins, Richard ;
Chatila, Raja ;
Herrera, Francisco .
INFORMATION FUSION, 2020, 58 :82-115
[6]  
Bastani O, 2018, ADV NEUR IN, V31
[7]   Explainable navigation system using fuzzy reinforcement learning [J].
Bautista-Montesano, Rolando ;
Bustamante-Bello, Rogelio ;
Ramirez-Mendoza, Ricardo A. .
INTERNATIONAL JOURNAL OF INTERACTIVE DESIGN AND MANUFACTURING - IJIDEM, 2020, 14 (04) :1411-1428
[8]   A Comprehensive Survey on the Application of Deep and Reinforcement Learning Approaches in Autonomous Driving [J].
Ben Elallid, Badr ;
Benamar, Nabil ;
Hafid, Abdelhakim Senhaji ;
Rachidi, Tajjeeddine ;
Mrani, Nabil .
JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (09) :7366-7390
[9]  
Biewald Lukas, 2020, Experiment tracking with weights and biases
[10]  
Breiman L., 1984, Classification and regression trees, DOI DOI 10.1201/9781315139470