A Reinforcement Learning Approach for Continuum Robot Control

被引:6
作者
Kargin, Turhan Can [1 ]
Kolota, Jakub [1 ]
机构
[1] Poznan Univ Tech, Inst Automat Control & Robot, Piotrowo 3A, PL-60965 Poznan, Poland
关键词
Reinforcement Learning; DDPG algorithm; Continuum robot; IMPLEMENTATION; KINEMATICS;
D O I
10.1007/s10846-023-02003-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Rigid joint manipulators are limited in their movement and degrees of freedom (DOF), while continuum robots possess a continuous backbone that allows for free movement and multiple DOF. Continuum robots move by bending over a section, taking inspiration from biological manipulators such as tentacles and trunks. This paper presents an implementation of a forward kinematics and velocity kinematics model to describe the planar continuum robot, along with the application of reinforcement learning (RL) as a control algorithm. In this paper, we have adopted the planar constant curvature representation for the forward kinematic modeling. This choice was made due to its straightforward implementation and its potential to fill the literature gap in the field RL-based control for planar continuum robots. The intended control mechanism is achieved through the use of Deep Deterministic Policy Gradient (DDPG), a RL algorithm that is suited for learning controls in continuous action spaces. After simulating the algorithm, it was observed that the planar continuum robot can autonomously move from any initial point to any desired goal point within the task space of the robot. By analyzing the results, we wanted to recommend a future direction for research in the field of continuum robot control, specifically in the application of RL algorithms. One potential area of focus could be the integration of sensory feedback, such as vision or force sensing, to improve the robot's ability to navigate complex environments. Additionally, exploring the use of different RL algorithms, such as Proximal Policy Optimization (PPO) or Trust Region Policy Optimization (TRPO), could lead to further advancements in the field. Overall, this paper demonstrates the potential for RL-based control of continuum robots and highlights the importance of continued research in this area.
引用
收藏
页数:14
相关论文
共 39 条
[1]  
Bailly Y, 2005, IEEE INT CONF ROBOT, P924
[2]   A New Soft Robot Control Method Using Model Predictive Control for a Pneumatically Actuated Humanoid [J].
Best, Charles M. ;
Gillespie, Morgan T. ;
Hyatt, Phillip ;
Rupert, Levi ;
Sherrod, Vallan ;
Killpack, Marc D. .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 2016, 23 (03) :75-84
[3]  
Brown N, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P5226
[4]   Continuum Robots for Medical Applications: A Survey [J].
Burgner-Kahrs, Jessica ;
Rucker, D. Caleb ;
Choset, Howie .
IEEE TRANSACTIONS ON ROBOTICS, 2015, 31 (06) :1261-1280
[5]  
de La Bourdonnaye F, 2019, J IEEE I C DEVELOP L, P93, DOI [10.1109/DEVLRN.2019.8850702, 10.1109/devlrn.2019.8850702]
[6]   Model-Based Control of Soft Robots A SURVEY OF THE STATE OF THE ART AND OPEN CHALLENGES [J].
Della Santina, Cosimo ;
Duriez, Christian ;
Rus, Daniela .
IEEE CONTROL SYSTEMS MAGAZINE, 2023, 43 (03) :30-65
[7]   Deep Direct Reinforcement Learning for Financial Signal Representation and Trading [J].
Deng, Yue ;
Bao, Feng ;
Kong, Youyong ;
Ren, Zhiquan ;
Dai, Qionghai .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (03) :653-664
[8]  
Francois-Lavet Vincent., 2017, Contributions to deep reinforcement learning and its applications in smartgrids
[9]  
Fujimoto S, 2018, Arxiv, DOI [arXiv:1802.09477, 10.48550/arXiv.1802.09477]
[10]  
Gandhi D, 2017, Arxiv, DOI arXiv:1704.05588