SAMBA: safe model-based & active reinforcement learning

被引:0
|
作者
Alexander I. Cowen-Rivers
Daniel Palenicek
Vincent Moens
Mohammed Amin Abdullah
Aivar Sootla
Jun Wang
Haitham Bou-Ammar
机构
[1] Huawei Noah’s Ark Lab,
[2] Technical University Darmstadt,undefined
[3] University College London,undefined
来源
Machine Learning | 2022年 / 111卷
关键词
Gaussian process; Safe reinforcement learning; Active learning;
D O I
暂无
中图分类号
学科分类号
摘要
In this paper, we propose SAMBA, a novel framework for safe reinforcement learning that combines aspects from probabilistic modelling, information theory, and statistics. Our method builds upon PILCO to enable active exploration using novel acquisition functions for out-of-sample Gaussian process evaluation optimised through a multi-objective problem that supports conditional-value-at-risk constraints. We evaluate our algorithm on a variety of safe dynamical system benchmarks involving both low and high-dimensional state representations. Our results show orders of magnitude reductions in samples and violations compared to state-of-the-art methods. Lastly, we provide intuition as to the effectiveness of the framework by a detailed analysis of our acquisition functions and safety constraints.
引用
收藏
页码:173 / 203
页数:30
相关论文
共 50 条
  • [1] SAMBA: safe model-based & active reinforcement learning
    Cowen-Rivers, Alexander, I
    Palenicek, Daniel
    Moens, Vincent
    Abdullah, Mohammed Amin
    Sootla, Aivar
    Wang, Jun
    Bou-Ammar, Haitham
    MACHINE LEARNING, 2022, 111 (01) : 173 - 203
  • [2] Safe Model-based Reinforcement Learning with Stability Guarantees
    Berkenkamp, Felix
    Turchetta, Matteo
    Schoellig, Angela P.
    Krause, Andreas
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [3] Safe Robot Execution in Model-Based Reinforcement Learning
    Martinez, David
    Alenya, Guillem
    Torras, Carme
    2015 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2015, : 6422 - 6427
  • [4] Conservative and Adaptive Penalty for Model-Based Safe Reinforcement Learning
    Ma, Yecheng Jason
    Shen, Andrew
    Bastani, Osbert
    Jayaraman, Dinesh
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 5404 - 5412
  • [5] Safe Model-Based Reinforcement Learning for Systems With Parametric Uncertainties
    Mahmud, S. M. Nahid
    Nivison, Scott A.
    Bell, Zachary I.
    Kamalapurkar, Rushikesh
    FRONTIERS IN ROBOTICS AND AI, 2021, 8
  • [6] Model-based Safe Reinforcement Learning using Variable Horizon Rollouts
    Gupta, Shourya
    Suryaman, Utkarsh
    Narava, Rahul
    Jha, Shashi Shekhar
    PROCEEDINGS OF 7TH JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MANAGEMENT OF DATA, CODS-COMAD 2024, 2024, : 100 - 108
  • [7] Provably Safe Model-Based Meta Reinforcement Learning: An Abstraction-Based Approach
    Sun, Xiaowu
    Fatnassi, Wael
    Santa Cruz, Ulices
    Shoukry, Yasser
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2963 - 2968
  • [8] Safe exploration in model-based reinforcement learning using control barrier functions
    Cohen, Max H.
    Belta, Calin
    AUTOMATICA, 2023, 147
  • [9] Safe Model-Based Reinforcement Learning With an Uncertainty-Aware Reachability Certificate
    Yu, Dongjie
    Zou, Wenjun
    Yang, Yujie
    Ma, Haitong
    Li, Shengbo Eben
    Yin, Yuming
    Chen, Jianyu
    Duan, Jingliang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (03) : 4129 - 4142
  • [10] Uncertainty-Aware Contact-Safe Model-Based Reinforcement Learning
    Kuo, Cheng-Yu
    Schaarschmidt, Andreas
    Cui, Yunduan
    Asfour, Tamim
    Matsubara, Takamitsu
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 3918 - 3925