Safe Reinforcement Learning Using Wasserstein Distributionally Robust MPC and Chance Constraint

被引:5
作者
Kordabad, Arash Bahari [1 ]
Wisniewski, Rafael [2 ]
Gros, Sebastien [1 ]
机构
[1] Norwegian Univ Sci & Technol NTNU, Dept Engn Cybernet, N-7034 Trondheim, Norway
[2] Aalborg Univ, Dept Elect Syst, DK-9220 Aalborg, Denmark
关键词
Safe reinforcement learning; model predictive control; distributionally robust optimization; chance constraint; conditional value at risk; Q-learning; OPTIMIZATION;
D O I
10.1109/ACCESS.2022.3228922
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we address the chance-constrained safe Reinforcement Learning (RL) problem using the function approximators based on Stochastic Model Predictive Control (SMPC) and Distributionally Robust Model Predictive Control (DRMPC). We use Conditional Value at Risk (CVaR) to measure the probability of constraint violation and safety. In order to provide a safe policy by construction, we first propose using parameterized nonlinear DRMPC at each time step. DRMPC optimizes a finite-horizon cost function subject to the worst-case constraint violation in an ambiguity set. We use a statistical ball around the empirical distribution with a radius measured by the Wasserstein metric as the ambiguity set. Unlike the sample average approximation SMPC, DRMPC provides a probabilistic guarantee of the out-of-sample risk and requires lower samples from the disturbance. Then the Q-learning method is used to optimize the parameters in the DRMPC to achieve the best closed-loop performance. Wheeled Mobile Robot (WMR) path planning with obstacle avoidance will be considered to illustrate the efficiency of the proposed method.
引用
收藏
页码:130058 / 130067
页数:10
相关论文
共 47 条