Data-Efficient Reinforcement Learning with Probabilistic Model Predictive Control

被引:0
作者
Kamthe, Sanket [1 ]
Deisenroth, Marc Peter [1 ]
机构
[1] Imperial Coll London, Dept Comp, London, England
来源
INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84 | 2018年 / 84卷
关键词
STABILITY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trial-and-error based reinforcement learning (RL) has seen rapid advancements in recent times, especially with the advent of deep neural networks. However, the majority of autonomous RL algorithms require a large number of interactions with the environment. A large number of interactions may be impractical in many real-world applications, such as robotics, and many practical systems have to obey limitations in the form of state space or control constraints. To reduce the number of system interactions while simultaneously handling constraints, we propose a model-based RL framework based on probabilistic Model Predictive Control (MPC). In particular, we propose to learn a probabilistic transition model using Gaussian Processes (GPs) to incorporate model uncertainty into long-term predictions, thereby, reducing the impact of model errors. We then use MPC to find a control sequence that minimises the expected long-term cost. We provide theoretical guarantees for first-order optimality in the GP-based transition models with deterministic approximate inference for long-term planning. We demonstrate that our approach does not only achieve state-of-the-art data efficiency, but also is a principled way for RL in constrained environments.
引用
收藏
页数:10
相关论文
共 50 条
[41]   Adaptive stochastic model predictive control via network ensemble learning [J].
Xiong, Weiliang ;
He, Defeng ;
Mu, Jianbin ;
Wang, Xiuli .
INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2023, 54 (16) :3013-3026
[42]   Stability of constrained model predictive control for networked control systems with data packet dropout [J].
Li, Zhijun ;
Wang, Lifeng ;
Lai, Xihuang ;
Xu, Shaohui .
2007 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS, VOLS 1-6, 2007, :3018-3023
[43]   Delay-aware model-based reinforcement learning for continuous control [J].
Chen, Baiming ;
Xu, Mengdi ;
Li, Liang ;
Zhao, Ding .
NEUROCOMPUTING, 2021, 450 :119-128
[44]   A Computationally Efficient Robust Model Predictive Control Framework for Uncertain Nonlinear Systems [J].
Koehler, Johannes ;
Soloperto, Raffaele ;
Mueller, Matthias A. ;
Allgoewer, Frank .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2021, 66 (02) :794-801
[45]   Dependable Data-based Design of Embedded Model Predictive Control [J].
Schmid, Patrick ;
Ebel, Henrik ;
Eberhard, Peter .
2022 EUROPEAN CONTROL CONFERENCE (ECC), 2022, :859-866
[46]   Model predictive control for constrained networked systems subject to data losses [J].
Franze, Giuseppe ;
Tedesco, Francesco ;
Famularo, Domenico .
AUTOMATICA, 2015, 54 :272-278
[47]   Nonlinear model predictive control from data: a set membership approach [J].
Canale, M. ;
Fagiano, L. ;
Signorile, M. C. .
INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2014, 24 (01) :123-139
[48]   Adaptive Control and Intersections with Reinforcement Learning [J].
Annaswamy, Anuradha M. .
ANNUAL REVIEW OF CONTROL ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 6 :65-93
[49]   Safety reinforcement learning control via transfer learning [J].
Zhang, Quanqi ;
Wu, Chengwei ;
Tian, Haoyu ;
Gao, Yabin ;
Yao, Weiran ;
Wu, Ligang .
AUTOMATICA, 2024, 166
[50]   Reinforcement Learning with Probabilistic Boolean Network Models of Smart Grid Devices [J].
Rivera Torres, Pedro Juan ;
Gershenson Garcia, Carlos ;
Sanchez Puig, Maria Fernanda ;
Kanaan Izquierdo, Samir .
COMPLEXITY, 2022, 2022