User-guided motion planning with reinforcement learning for human-robot collaboration in smart manufacturing

被引:11
作者
Yu, Tian [1 ]
Chang, Qing [1 ]
机构
[1] Univ Virginia, Dept Mech & Aerosp Engn, Charlottesville, VA 22904 USA
基金
美国国家科学基金会;
关键词
Human -robot collaboration; Learning from demonstration; Motion planning; Reinforcement learning;
D O I
10.1016/j.eswa.2022.118291
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In today's manufacturing system, robots are expected to perform increasingly complex manipulation tasks in collaboration with humans. However, current industrial robots are still largely preprogrammed with very little autonomy and still required to be reprogramed by robotics experts for even slightly changed tasks. Therefore, it is highly desirable that robots can adapt to certain task changes with motion planning strategies to easily work with non-robotic experts in manufacturing environments. In this paper, we propose a user-guided motion planning algorithm in combination with reinforcement learning (RL) method to enable robots automatically generate their motion plans for new tasks by learning from a few kinesthetic human demonstrations. Features of common human demonstrated tasks in a specific application environment, e.g., desk assembly or warehouse loading/ unloading are abstracted and saved in a library. The definition of semantical similarity between features in the library and features of a new task is proposed and further used to construct the reward function in RL. To achieve an adaptive motion plan facing task changes or new task requirements, features embedded in the library are mapped to appropriate task segments based on the trained motion planning policy using Q-learning. A new task can be either learned as a combination of a few features in the library or a requirement for further human demonstration if the current library is insufficient for the new task. We evaluate our approach on a 6 DOF UR5e robot on multiple tasks and scenarios and show the effectiveness of our method with respect to different scenarios.
引用
收藏
页数:13
相关论文
共 32 条
  • [1] Robust trajectory learning and approximation for robot programming by demonstration
    Aleotti, J.
    Caselli, S.
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2006, 54 (05) : 409 - 413
  • [2] A survey of robot learning from demonstration
    Argall, Brenna D.
    Chernova, Sonia
    Veloso, Manuela
    Browning, Brett
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2009, 57 (05) : 469 - 483
  • [3] Berenson D., 2009, IEEE INT C ROB AUT, P625
  • [4] Bovet D. P., 1994, INTRO THEORY COMPLEX, V7
  • [5] On learning, representing, and generalizing a task in a humanoid robot
    Calinon, Sylvain
    Guenter, Florent
    Billard, Aude
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2007, 37 (02): : 286 - 298
  • [6] Cobot programming for collaborative industrial tasks: An overview
    El Zaatari, Shirine
    Marei, Mohamed
    Li, Weidong
    Usman, Zahid
    [J]. ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 116 : 162 - 180
  • [7] Figueredo L. F. da C., 2016, KINEMATIC CONTROL BA
  • [8] Control of Redundant Robots Under Hard Joint Constraints: Saturation in the Null Space
    Flacco, Fabrizio
    De Luca, Alessandro
    Khatib, Oussama
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2015, 31 (03) : 637 - 654
  • [9] Learning Non-linear Multivariate Dynamics of Motion in Robotic Manipulators
    Gribovskaya, E.
    Khansari-Zadeh, S. M.
    Billard, A.
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2011, 30 (01) : 80 - 117
  • [10] Randomized multi-modal motion planning for a humanoid robot manipulation task
    Hauser, Kris
    Ng-Thow-Hing, Victor
    [J]. INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2011, 30 (06) : 678 - 698