Safe Reinforcement Learning of Robot Trajectories in the Presence of Moving Obstacles

被引:0
|
作者
Kiemel, Jonas [1 ]
Righetti, Ludovic [2 ]
Kroeger, Torsten [1 ]
Asfour, Tamim [1 ]
机构
[1] Karlsruhe Inst Technol KIT, Inst Anthropomat & Robot IAR, D-76131 Karlsruhe, Germany
[2] NYU, Tandon Sch Engn, Brooklyn, NY 11201 USA
来源
IEEE ROBOTICS AND AUTOMATION LETTERS | 2024年 / 9卷 / 12期
关键词
Robots; Collision avoidance; Trajectory; Safety; Stochastic processes; Reinforcement learning; Training; Real-time systems; Quadrotors; Kinematics; Motion control; reinforcement learning; robot safety; collision avoidance;
D O I
10.1109/LRA.2024.3488402
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In this paper, we present an approach for learning collision-free robot trajectories in the presence of moving obstacles. As a first step, we train a backup policy to generate evasive movements from arbitrary initial robot states using model-free reinforcement learning. When learning policies for other tasks, the backup policy can be used to estimate the potential risk of a collision and to offer an alternative action if the estimated risk is considered too high. No matter which action is selected, our action space ensures that the kinematic limits of the robot joints are not violated. We analyze and evaluate two different methods for estimating the risk of a collision. A physics simulation performed in the background is computationally expensive but provides the best results in deterministic environments. If a data-based risk estimator is used instead, the computational effort is significantly reduced, but an additional source of error is introduced. For evaluation, we successfully learn a reaching task and a basketball task while keeping the risk of collisions low. The results demonstrate the effectiveness of our approach for deterministic and stochastic environments, including a human-robot scenario and a ball environment, where no state can be considered permanently safe. By conducting experiments with a real robot, we show that our approach can generate safe trajectories in real time.
引用
收藏
页码:11353 / 11360
页数:8
相关论文
共 50 条
  • [31] A REDUCED-ORDER ANALYTICAL SOLUTION TO MOBILE ROBOT TRAJECTORY GENERATION IN THE PRESENCE OF MOVING OBSTACLES
    Wang, J.
    Qu, Z.
    Guo, Y.
    Yang, J.
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2009, 24 (04): : 283 - 291
  • [32] Reinforcement learning method for target hunting control of multi-robot systems with obstacles
    Fan, Zhilin
    Yang, Hongyong
    Liu, Fei
    Liu, Li
    Han, Yilin
    INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (12) : 11275 - 11298
  • [33] Autonomous Robot Navigation with Self-learning for Collision Avoidance with Randomly Moving Obstacles
    Zhang, Yunfei
    de Silva, Clarence W.
    Su, Dijia
    Xue, Youtai
    2014 PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE & EDUCATION (ICCSE 2014), 2014, : 117 - 122
  • [34] A human-centered safe robot reinforcement learning framework with interactive behaviors
    Gu, Shangding
    Kshirsagar, Alap
    Du, Yali
    Chen, Guang
    Peters, Jan
    Knoll, Alois
    FRONTIERS IN NEUROROBOTICS, 2023, 17
  • [35] Towards Safe Human-Robot Collaboration Using Deep Reinforcement Learning
    El-Shamouty, Mohamed
    Wu, Xinyang
    Yang, Shanqi
    Albus, Marcel
    Huber, Marco F.
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 4899 - 4905
  • [36] Reinforcement Learning for Safe Robot Control using Control Lyapunov Barrier Functions
    Du, Desong
    Han, Shaohang
    Qi, Naiming
    Ammar, Haitham Bou
    Wang, Jun
    Pan, Wei
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 9442 - 9448
  • [37] Safe multi-agent reinforcement learning for multi-robot control
    Gu, Shangding
    Kuba, Jakub Grudzien
    Chen, Yuanpei
    Du, Yali
    Yang, Long
    Knoll, Alois
    Yang, Yaodong
    ARTIFICIAL INTELLIGENCE, 2023, 319
  • [38] On Normative Reinforcement Learning via Safe Reinforcement Learning
    Neufeld, Emery A.
    Bartocci, Ezio
    Ciabattoni, Agata
    PRIMA 2022: PRINCIPLES AND PRACTICE OF MULTI-AGENT SYSTEMS, 2023, 13753 : 72 - 89
  • [39] Avoiding Moving Obstacles with Stochastic Hybrid Dynamics using PEARL: PrEference Appraisal Reinforcement Learning
    Faust, Aleksandra
    Chiang, Hao-Tien
    Rackley, Nathanael
    Tapia, Lydia
    2016 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2016, : 484 - 490
  • [40] Tracking Control of an Underwater Robot in the Presence of Obstacles
    Moosavian, S. Ali A.
    Khalaji, Ali Keymasi
    Tabataba'i-Nasab, Fahimeh S.
    2017 5TH RSI INTERNATIONAL CONFERENCE ON ROBOTICS AND MECHATRONICS (ICROM 2017), 2017, : 298 - 303