A Safe Reinforcement Learning driven Weights-varying Model Predictive Control for Autonomous Vehicle Motion Control

被引:0
作者
Zarrouki, Baha [1 ,2 ]
Spanakakis, Marios
Betz, Johannes
机构
[1] Tech Univ Munich, TUM Sch Engn & Design, Automot Technol, Munich, Germany
[2] Tech Univ Munich, TUM Sch Engn & Design, Autonomous Vehicle Syst, Munich, Germany
来源
2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024 | 2024年
关键词
MPC;
D O I
10.1109/IV55156.2024.10588747
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Determining the optimal cost function parameters of Model Predictive Control (MPC) to optimize multiple control objectives is a challenging and time-consuming task. Multi-objective Bayesian Optimization (BO) techniques solve this problem by determining a Pareto optimal parameter set for an MPC with static weights. However, a single parameter set may not deliver the most optimal closed-loop control performance when the context of the MPC operating conditions changes during its operation, urging the need to adapt the cost function weights at runtime. Deep Reinforcement Learning (RL) algorithms can automatically learn context-dependent optimal parameter sets and dynamically adapt for a Weights-varying MPC (WMPC). However, learning cost function weights from scratch in a continuous action space may lead to unsafe operating states. To solve this, we propose a novel approach limiting the RL action space within a safe learning space that we represent by a catalog of pre-optimized feasible BO Pareto-optimal weight sets. We conceive an RL agent not to learn in a continuous space but to select the most optimal discrete actions, each corresponding to a single set of Pareto optimal weights, by proactively anticipating upcoming control tasks in a context-dependent manner. This approach introduces a two-step optimization: (1) safety-critical with BO and (2) performance-driven with RL. Hence, even an untrained RL agent guarantees a safe and optimal performance. Simulation results demonstrate that an untrained RL-WMPC shows Pareto-optimal closed-loop behavior and training the RL-WMPC helps exhibit a performance beyond the Pareto-front. The code used in this research is publicly accessible as open-source software: https://github.com/bzarr/TUM-CONTROL
引用
收藏
页码:1401 / 1408
页数:8
相关论文
共 50 条
[41]   Experimental validation of model predictive control stability for autonomous driving [J].
Lima, Pedro F. ;
Pereira, Goncalo Collares ;
Martensson, Jonas ;
Wahlberg, Bo .
CONTROL ENGINEERING PRACTICE, 2018, 81 :244-255
[42]   A Model Predictive Control Approach of Optimal Autonomous Laboratory Management [J].
Zhao, Zhihong ;
Lin, Xiaotian ;
Rodriguez-Andina, Juan J. ;
Li, Zhengkai .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2025,
[43]   The Design of Performance Guaranteed Autonomous Vehicle Control for Optimal Motion in Unsignalized Intersections [J].
Nemeth, Balazs ;
Gaspar, Peter .
APPLIED SCIENCES-BASEL, 2021, 11 (08)
[44]   Integrated Spatial Kinematics-Dynamics Model Predictive Control for Collision-Free Autonomous Vehicle Tracking [J].
Yang, Weishan ;
Su, Yixin ;
Chen, Yuepeng ;
Lian, Cheng .
ACTUATORS, 2024, 13 (04)
[45]   Nonlinear Model Predictive Control for Time-Optimal Turning Around of an Autonomous Vehicle Under Steering Lag [J].
Li, Hongbo ;
Li, Bowen ;
Yang, Hongjiu ;
Mu, Chaoxu .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2025, 30 (01) :577-586
[46]   Adaptive Model Predictive Control for Constrained Time Varying Systems [J].
Tanaskovic, Marko ;
Fagiano, Lorenzo ;
Gligorovski, Vojislav .
2018 EUROPEAN CONTROL CONFERENCE (ECC), 2018, :1698-1703
[47]   Robust Model Predictive Control with Time-varying Tubes [J].
Bumroongsri, Pornchai ;
Kheawhom, Soorathep .
INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2017, 15 (04) :1479-1484
[48]   Predictive Control of Voltage Source Inverter: An Online Reinforcement Learning Solution [J].
Liu, Xing ;
Qiu, Lin ;
Fang, Youtong ;
Wang, Kui ;
Li, Yongdong ;
Rodriguez, Jose .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (07) :6591-6600
[49]   Obstacle parameter modeling for model predictive control of the unmanned vehicle [J].
Yeu, Jung-Yun ;
Kim, Woo-Hyun ;
Im, Jun-Hyuck ;
Lee, Dal-Ho ;
Jee, Gyu-In .
Journal of Institute of Control, Robotics and Systems, 2012, 18 (12) :1132-1138
[50]   Model Predictive Control for Unmanned Tracked Vehicle Path Following [J].
Yan, Hao ;
Yang, Fuwei ;
Wang, Zhixiang .
2018 INTERNATIONAL CONFERENCE ON SENSING, DIAGNOSTICS, PROGNOSTICS, AND CONTROL (SDPC), 2018, :101-106