A Vision-Based Bio-Inspired Reinforcement Learning Algorithms for Manipulator Obstacle Avoidance

被引:3
作者
Singh, Abhilasha [1 ]
Shakeel, Mohamed [2 ]
Kalaichelvi, V [1 ]
Karthikeyan, R. [2 ]
机构
[1] Birla Inst Technol & Sci Pilani, Dept Elect & Elect Engn, Dubai Campus,POB 345 055, Dubai, U Arab Emirates
[2] Birla Inst Technol & Sci Pilani, Dept Mech Engn, Dubai Campus,POB 345 055, Dubai, U Arab Emirates
关键词
Q-learning; DQN; SARSA; DDQN; homogeneous transformation; optimization; obstacle avoidance; MOBILE ROBOT; ENVIRONMENTS;
D O I
10.3390/electronics11213636
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Path planning for robotic manipulators has proven to be a challenging issue in industrial applications. Despite providing precise waypoints, the traditional path planning algorithm requires a predefined map and is ineffective in complex, unknown environments. Reinforcement learning techniques can be used in cases where there is a no environmental map. For vision-based path planning and obstacle avoidance in assembly line operations, this study introduces various Reinforcement Learning (RL) algorithms based on discrete state-action space, such as Q-Learning, Deep Q Network (DQN), State-Action-Reward-State-Action (SARSA), and Double Deep Q Network (DDQN). By positioning the camera in an eye-to-hand position, this work used color-based segmentation to identify the locations of obstacles, start, and goal points. The homogeneous transformation technique was used to further convert the pixel values into robot coordinates. Furthermore, by adjusting the number of episodes, steps per episode, learning rate, and discount factor, a performance study of several RL algorithms was carried out. To further tune the training hyperparameters, genetic algorithms (GA) and particle swarm optimization (PSO) were employed. The length of the path travelled, the average reward, the average number of steps, and the time required to reach the objective point were all measured and compared for each of the test cases. Finally, the suggested methodology was evaluated using a live camera that recorded the robot workspace in real-time. The ideal path was then drawn using a TAL BRABO 5 DOF manipulator. It was concluded that waypoints obtained via Double DQN showed an improved performance and were able to avoid the obstacles and reach the goal point smoothly and efficiently.
引用
收藏
页数:26
相关论文
共 38 条
[1]   A Novel Hybrid Path Planning Method Based on Q-Learning and Neural Network for Robot Arm [J].
Abdi, Ali ;
Adhikari, Dibash ;
Park, Ju Hong .
APPLIED SCIENCES-BASEL, 2021, 11 (15)
[2]  
Abed M.S., 2021, ENG TECHNOL J, V39, P820, DOI [10.30684/etj.v39i5A.1941, DOI 10.30684/ETJ.V39I5A.1941]
[3]  
Attamimi M, 2021, 2021 8TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTERSCIENCE AND INFORMATICS (EECSI) 2021, P169, DOI 10.23919/EECSI53397.2021.9624254
[4]   An Overview of Nature-Inspired, Conventional, and Hybrid Methods of Autonomous Vehicle Path Planning [J].
Ayawli, Ben Beklisi Kwame ;
Chellali, Ryad ;
Appiah, Albert Yaw ;
Kyeremeh, Frimpong .
JOURNAL OF ADVANCED TRANSPORTATION, 2018,
[5]   Modeling Method of Autonomous Robot Manipulator Based on D-H Algorithm [J].
Cai, Jie ;
Deng, Jinlian ;
Zhang, Wei ;
Zhao, Weisheng .
MOBILE INFORMATION SYSTEMS, 2021, 2021
[6]  
Deshpande V., 2014, INT J ROBOT RES, V4, P17
[7]  
Dong YS, 2020, INT CONF SOFTW ENG, P52, DOI [10.1109/ICISCE50968.2020.00021, 10.1109/icsess49938.2020.9237641, 10.1109/ICSESS49938.2020.9237641]
[8]   Reinforcement Learning for Position Control Problem of a Mobile Robot [J].
Farias, Gonzalo ;
Garcia, Gonzalo ;
Montenegro, Guelis ;
Fabregas, Ernesto ;
Dormido-Canto, Sebastian ;
Dormido, Sebastian .
IEEE ACCESS, 2020, 8 :152941-152951
[9]   Deep Reinforcement Learning for Indoor Mobile Robot Path Planning [J].
Gao, Junli ;
Ye, Weijie ;
Guo, Jing ;
Li, Zhongjuan .
SENSORS, 2020, 20 (19) :1-15
[10]  
Hachour O., 2008, Int. J. Syst. Appl. Eng. Dev, V2, P178