Using Goal-Conditioned Reinforcement Learning With Deep Imitation to Control Robot Arm in Flexible Flat Cable Assembly Task

被引:5
|
作者
Li, Jingchen [1 ]
Shi, Haobin [1 ]
Hwang, Kao-Shing [2 ,3 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Peoples R China
[2] Natl Sun Yat Sen Univ, Dept Elect Engn, Kaohsiung 81164, Taiwan
[3] Kaohsiung Med Univ, Dept Healthcare Adm & Med Informat, Kaohsiung 80708, Taiwan
基金
中国国家自然科学基金;
关键词
Robots; Manipulators; Reinforcement learning; Task analysis; Connectors; Service robots; Production; Deep reinforcement learning; robot arm; intelligent assembly;
D O I
10.1109/TASE.2023.3323307
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Leveraging reinforcement learning on high-precision decision-making in Robot Arm assembly scenes is a desired goal in the industrial community. However, tasks like Flexible Flat Cable (FFC) assembly, which require highly trained workers, pose significant challenges due to sparse rewards and limited learning conditions. In this work, we propose a goal-conditioned self-imitation reinforcement learning method for FFC assembly without relying on a specific end-effector, where both perception and behavior plannings are learned through reinforcement learning. We analyze the challenges faced by Robot Arm in high-precision assembly scenarios and balance the breadth and depth of exploration during training. Our end-to-end model consists of hindsight and self-imitation modules, allowing the Robot Arm to leverage futile exploration and optimize successful trajectories. Our method does not require rule-based or manual rewards, and it enables the Robot Arm to quickly find feasible solutions through experience relabeling, while unnecessary explorations are avoided. We train the FFC assembly policy in a simulation environment and transfer it to the real scenario by using domain adaptation. We explore various combinations of hindsight and self-imitation learning, and discuss the results comprehensively. Experimental findings demonstrate that our model achieves fast and advanced flexible flat cable assembly, surpassing other reinforcement learning-based methods.Note to Practitioners-The motivation of this article stems from the need to develop an efficient and accurate FFC assembly policy for 3C (Computer, Communication, and Consumer Electronic) industry, promoting the development of intelligent manufacturing. Traditional control methods are incompetent to complete such a high-precision task with Robot Arm due to the difficult-to-model connectors, and existing reinforcement learning methods cannot converge with restricted epochs because of the difficult goals or trajectories. To quickly learn a high-quality assembly for Robot Arm and accelerate the convergence speed, we combine the goal-conditioned reinforcement learning and self-imitation mechanism, balancing the depth and breadth of exploration. The proposal takes visual information and six-dimensions force as state, obtaining satisfactory assembly policies. We build a simulation scene by the Pybullet platform and pre-train the Robot Arm on it, and then the pre-trained policies can be reused in real scenarios with finetuning.
引用
收藏
页码:6217 / 6228
页数:12
相关论文
共 50 条
  • [41] Dynamic warning zone and a short-distance goal for autonomous robot navigation using deep reinforcement learning
    Montero, Estrella Elvia
    Mutahira, Husna
    Pico, Nabih
    Muhammad, Mannan Saeed
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (01) : 1149 - 1166
  • [42] Scalable Multi-Robot Task Allocation Using Graph Deep Reinforcement Learning with Graph Normalization
    Zhang, Zhenqiang
    Jiang, Xiangyuan
    Yang, Zhenfa
    Ma, Sile
    Chen, Jiyang
    Sun, Wenxu
    ELECTRONICS, 2024, 13 (08)
  • [43] Deep reinforcement learning and decoupling proportional-integral-derivative control of a humanoid cable-driven hybrid robot
    Liu, Yong
    Luo, Zhisheng
    Qian, Sen
    Wang, Shuaikang
    Wu, Zhe
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2024, 21 (03):
  • [44] Brain-Machine Interface Control of a Robot Arm using Actor-Critic Reinforcement Learning
    Pohlmeyer, Eric A.
    Mahmoudi, Babak
    Geng, Shijia
    Prins, Noeine
    Sanchez, Justin C.
    2012 ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2012, : 4108 - 4111
  • [45] Sim-Real Mapping of an Image-Based Robot Arm Controller Using Deep Reinforcement Learning
    Sasaki, Minoru
    Muguro, Joseph
    Kitano, Fumiya
    Njeri, Waweru
    Matsushita, Kojiro
    APPLIED SCIENCES-BASEL, 2022, 12 (20):
  • [46] Motion Control of Non-Holonomic Constrained Mobile Robot Using Deep Reinforcement Learning
    Gao, Rui
    Gao, Xueshan
    Liang, Peng
    Han, Feng
    Lan, Bingqing
    Li, Jingye
    Li, Jian
    Li, Simin
    2019 IEEE 4TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS (ICARM 2019), 2019, : 348 - 353
  • [47] Safety-based Dynamic Task Offloading for Human-Robot Collaboration using Deep Reinforcement Learning
    Ruggeri, Franco
    Terra, Ahmad
    Hata, Alberto
    Inam, Rafia
    Leite, Iolanda
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 2119 - 2126
  • [48] Adaptive Flexible Switching Mode Control of the Aircraft Skin Inspection Robot Using Integral Reinforcement Learning
    Wu, Xuewei
    Wang, Congqing
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 2220 - 2225
  • [49] Open-Loop Motion Control of a Hydraulic Soft Robotic Arm Using Deep Reinforcement Learning
    Zhang, Yunce
    Wang, Tao
    Tan, Ning
    Zhu, Shiqiang
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2021, PT I, 2021, 13013 : 302 - 312
  • [50] Dynamic Task Planning for Multi-Arm Harvesting Robots Under Multiple Constraints Using Deep Reinforcement Learning
    Xie, Feng
    Guo, Zhengwei
    Li, Tao
    Feng, Qingchun
    Zhao, Chunjiang
    HORTICULTURAE, 2025, 11 (01)