Deep reinforcement learning based proactive dynamic obstacle avoidance for safe human-robot collaboration

被引:1
作者
Xia, Wanqing [1 ]
Lu, Yuqian [1 ]
Xu, Weiliang [1 ]
Xu, Xun [1 ]
机构
[1] Univ Auckland, 20 Symond St, Auckland 1010, New Zealand
关键词
Human-robot collaboration; Dynamic obstacle avoidance; Deep reinforcement learning; Reward engineering;
D O I
10.1016/j.mfglet.2024.09.151
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Ensuring the health and safety of human operators is paramount in manufacturing, particularly in human-robot collaborative environments. In this paper, we present a deep reinforcement learning-based trajectory planning method for a robotic manipulator designed to avoid collisions with human body parts in real-time while achieving its goal. We modelled the human arm as a freely moving cylinder in 3D space and formulated the dynamic obstacle avoidance problem as a Markov decision process. The algorithm was tested in a simulated environment that closely mimics our laboratory environment, with the goal of training a deep reinforcement learning model for autonomous task completion. A composite reward function was developed to balance the effects of different environmental variables, and the soft-actor critic algorithm was employed. The trained model demonstrated a 93% success rate in avoiding dynamic obstacles while achieving its goals when tested on a generated data set. (c) 2024 The Authors. Published by ELSEVIER Ltd. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/)
引用
收藏
页码:1246 / 1256
页数:11
相关论文
共 33 条
  • [1] Casalino A, 2019, IEEE INT CONF ROBOT, P6540, DOI [10.1109/ICRA.2019.8793847, 10.1109/icra.2019.8793847]
  • [2] Deep Reinforcement Learning Based Trajectory Planning Under Uncertain Constraints
    Chen, Lienhung
    Jiang, Zhongliang
    Cheng, Long
    Knoll, Alois C.
    Zhou, Mingchuan
    [J]. FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [3] A deep reinforcement learning based method for real-time path planning and dynamic obstacle avoidance
    Chen, Pengzhan
    Pei, Jiean
    Lu, Weiqing
    Li, Mingzhen
    [J]. NEUROCOMPUTING, 2022, 497 : 64 - 75
  • [4] Cheng X., 2022, 2022 IEEE 11 DAT DRI, P1136
  • [5] El-Shamouty M, 2020, IEEE INT CONF ROBOT, P4899, DOI [10.1109/icra40945.2020.9196924, 10.1109/ICRA40945.2020.9196924]
  • [6] Fujimoto S, 2018, PR MACH LEARN RES, V80
  • [7] Gu SX, 2016, PR MACH LEARN RES, V48
  • [8] Haarnoja T, 2018, PR MACH LEARN RES, V80
  • [9] Haddadin S., 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2010), P3109, DOI 10.1109/IROS.2010.5650246
  • [10] Hadfield-Menell D, 2017, ADV NEUR IN, V30