Event-Triggered Model Predictive Control With Deep Reinforcement Learning for Autonomous Driving

被引:21
作者
Dang, Fengying [1 ]
Chen, Dong [2 ]
Chen, Jun [3 ]
Li, Zhaojian [4 ]
机构
[1] Univ Michigan, Transportat Res Inst, Ann Arbor, MI 48109 USA
[2] Michigan State Univ, Dept Elect & Comp Engn, E Lansing, MI 48824 USA
[3] Oakland Univ, Dept Elect & Comp Engn, Rochester, MI 48309 USA
[4] Michigan State Univ, Dept Mech Engn, E Lansing, MI 48824 USA
来源
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES | 2024年 / 9卷 / 01期
基金
美国国家科学基金会;
关键词
Autonomous vehicles; Optimal control; Vehicle dynamics; Predictive control; Computational modeling; Tires; Q-learning; double Q-learning (DDQN); event-triggered model predictive control (eMPC); proximal policy optimization (PPO); reinforcement learning (RL); soft actor-critic (SAC);
D O I
10.1109/TIV.2023.3329785
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event-triggered model predictive control (eMPC) is a popular optimal control method with an aim to alleviate the computation and/or communication burden of MPC. However, it generally requires a priori knowledge of the closed-loop system behavior along with the communication characteristics for designing the event-trigger policy. This paper attempts to solve this challenge by proposing an efficient eMPC framework and demonstrates successful implementation of this framework on the autonomous vehicle path following. First of all, a model-free reinforcement learning (RL) agent is used to learn the optimal event-trigger policy without the need for a complete dynamical system and communication knowledge in this framework. Furthermore, techniques including prioritized experience replay (PER) buffer and long short-term memory (LSTM) are employed to foster exploration and improve training efficiency. In this paper, we use the proposed framework with three deep RL algorithms, i.e., Double Q-learning (DDQN), Proximal Policy Optimization (PPO), and Soft Actor-Critic (SAC), to solve this problem. Results show that all three deep RL-based eMPC (deep-RL-eMPC) can achieve better evaluation performance than the conventional threshold-based and previous linear Q-based approach in the autonomous path following. In particular, PPO-eMPC with LSTM and DDQN-eMPC with PER and LSTM obtain a superior balance between the closed-loop control performance and event-trigger frequency.
引用
收藏
页码:459 / 468
页数:10
相关论文
共 57 条
[1]   Communication-Constrained Active Suspension Control for Networked In-Wheel Motor-Driven Electric Vehicles With Dynamic Dampers [J].
Ahmad, Iftikhar ;
Ge, Xiaohua ;
Han, Qing-Long .
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2022, 7 (03) :590-602
[2]   An Overview of Nature-Inspired, Conventional, and Hybrid Methods of Autonomous Vehicle Path Planning [J].
Ayawli, Ben Beklisi Kwame ;
Chellali, Ryad ;
Appiah, Albert Yaw ;
Kyeremeh, Frimpong .
JOURNAL OF ADVANCED TRANSPORTATION, 2018,
[3]  
Baumann D, 2018, IEEE DECIS CONTR P, P943, DOI 10.1109/CDC.2018.8619335
[4]   Robust Event-Triggered MPC With Guaranteed Asymptotic Bound and Average Sampling Rate [J].
Brunner, Florian David ;
Heemels, W. P. M. H. ;
Allgower, Frank .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2017, 62 (11) :5694-5709
[5]  
Chen J, 2022, P AMER CONTR CONF, P3342, DOI 10.23919/ACC53348.2022.9867347
[6]   Comparison of Event-Triggered Model Predictive Control for Autonomous Vehicle Path Tracking [J].
Chen, Jun ;
Yi, Zonggen .
5TH IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (IEEE CCTA 2021), 2021, :808-813
[7]  
Christodoulou P, 2019, Arxiv, DOI arXiv:1910.07207
[8]   Model predictive control and its application in agriculture: A review [J].
Ding, Ying ;
Wang, Liang ;
Li, Yongwei ;
Li, Daoliang .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2018, 151 :104-117
[9]  
Dosovitskiy A., 2017, P 1 ANN C ROBOT LEAR, P1, DOI DOI 10.48550/ARXIV.1711.03938
[10]  
Eqtami A, 2011, IEEE DECIS CONTR P, P3392, DOI 10.1109/CDC.2011.6161348