Model-Free Reinforcement Learning with Ensemble for a Soft Continuum Robot Arm

被引:34
作者
Morimoto, Ryota [1 ]
Nishikawa, Satoshi [1 ]
Niiyama, Ryuma [1 ]
Kuniyoshi, Yasuo [1 ]
机构
[1] Univ Tokyo, Grad Sch Informat Sci & Technol, Bunkyo Ku, 7-3-1 Hongo, Tokyo, Japan
来源
2021 IEEE 4TH INTERNATIONAL CONFERENCE ON SOFT ROBOTICS (ROBOSOFT) | 2021年
关键词
KINEMATICS;
D O I
10.1109/RoboSoft51838.2021.9479340
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Soft robots have more passive degrees of freedom (DoFs) than rigid-body robots, which makes controller design difficult. Model-free reinforcement learning (RL) is a promising tool to resolve control problems in soft robotics alongside detailed and elaborate modeling. However, the adaptation of RL to soft robots requires consideration of the unique nature of soft bodies. In this work, a continuum robot arm is used as an example of a soft robot, and we propose an Ensembled Lightweight model-Free reinforcement learning Network (ELFNet), which is an RL framework with a computationally light ensemble. We demonstrated that the proposed system could learn control policies for a continuum robot arm to reach target positions using its tip not only in simulations but also in the real world. We used a pneumatically controlled continuum robot arm that operates with nine flexible rubber artificial muscles. Each artificial muscle can be controlled independently by pressure control valves, demonstrating that the policy can be learned using a real robot alone. We found that our method is more suitable for compliant robots than other RL methods because the sample efficiency is better than that of the other methods, and there is a significant difference in the performance when the number of passive DoFs is large. This study is expected to lead to the development of model-free RL in future soft robot control.
引用
收藏
页码:141 / 148
页数:8
相关论文
共 37 条
[1]   Deep Reinforcement Learning for Soft, Flexible Robots: Brief Review with Impending Challenges [J].
Bhagat, Sarthak ;
Banerjee, Hritwick ;
Tse, Zion Tsz Ho ;
Ren, Hongliang .
ROBOTICS, 2019, 8 (01)
[2]  
Brockman Greg, 2016, arXiv
[3]  
Bruder D, 2019, IEEE INT CONF ROBOT, P6244, DOI [10.1109/ICRA.2019.8793766, 10.1109/icra.2019.8793766]
[4]  
Buchler D, 2020, Arxiv, DOI arXiv:2006.05935
[5]  
Chattopadhyay S, 2018, 2018 INTERNATIONAL CONFERENCE ON CONTROL, POWER, COMMUNICATION AND COMPUTING TECHNOLOGIES (ICCPCCT), P222, DOI 10.1109/ICCPCCT.2018.8574225
[6]  
Chen Yang, 2019, 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), P3019, DOI 10.1109/ROBIO49542.2019.8961852
[7]   Kinematic Calibration of a Multisection Bionic Manipulator [J].
Escande, Coralie ;
Chettibi, Taha ;
Merzouki, Rochdi ;
Coelen, Vincent ;
Pathak, Pushparaj Mani .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2015, 20 (02) :663-674
[8]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[9]   A Fuzzy Reinforcement Learning Approach for Continuum Robot Control [J].
Goharimanesh, M. ;
Mehrkish, A. ;
Janabi-Sharifi, F. .
JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2020, 100 (3-4) :809-826
[10]  
Haarnoja T, 2018, Arxiv, DOI [arXiv:1801.01290, DOI 10.48550/ARXIV.1801.01290]