Collaborative obstacle avoidance trajectory planning for mobile robotic arms based on artificial potential field DDPG algorithm

被引:0
|
作者
Li, Yong [1 ]
Zhang, Chaoxing [1 ]
Chai, Liaoning [1 ]
机构
[1] Key Laboratory of Industrial Internet of Things and Network Control, Ministry of Education, Chongqing University of Posts and Telecommunications, Chongqing
来源
Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS | 2024年 / 30卷 / 12期
关键词
artificial potential field; deep deterministic policy gradient; guided training; mobile robotic arm; obstacle avoidance trajectory planning;
D O I
10.13196/j.cims.2023.0369
中图分类号
学科分类号
摘要
To improve the obstacle avoidance trajectory planning ability of mobile robotic arm in narrow channel and obstacle constraint situations, by combining Artificial Potential Field method (APF) and Deep Deterministic Policy Gradient algorithm (DDPG), an improved algorithm named APF-DDPG was proposed.The APF planning was designed for the robotic arm to get the approximate pose, and the research problem was represented as a Markov decision process. The state space, action space and reward and punishment functions were designed, and the planning process was analyzed and processed in phases. A mechanism for guiding was designed to transition the various control phases, which the obstacle avoidance phase of the training was dominated by DDPG, and the approximate pose dominated the goal planning phase to guide the DDPG for the training. Thus the strategy model for planning was obtained from the training. Finally, simulation experiments of fixed and random state scenarios were established and designed to verify the effectiveness of the proposed algorithm. The experimental results showed that APF-DDPG algorithm could be trained with higher convergence efficiency to obtain a policy model with more efficient control performance by comparing with the traditional DDPG algorithm. © 2024 CIMS. All rights reserved.
引用
收藏
页码:4282 / 4291
页数:9
相关论文
共 20 条
  • [1] LIU Ronghai, YUAN Hui, YANG Yingchun, Et al., Research status and development direction of mobile manipulator, Tool Engineering, 51, 5, pp. 3-8, (2017)
  • [2] CHEN Chunzhao, LIU Yi, WANG Geng, Obstacle avoidance path planning of mobile manipulator based on ACO-RRT algorithm, Journal of Henan Polytechnic University: Natural Science, 42, 3, pp. 95-102, (2023)
  • [3] AVINESH P, BIBHYA S, JITO V, Et al., Motion control of an articulated mobile manipulator in 3D using the Lyapunov-based control scheme, International Journal of Control, 95, 9, pp. 2581-2595, (2022)
  • [4] CHEN Jiapan, ZHENG Minhua, Research on robot operation behavior based on deep reinforcement learning, Robot, 44, 2, pp. 236-256, (2022)
  • [5] ZHANG Haoyu, XIONG Kai, Research on gait control of quadruped robot based on proximal strategy optimization algorithm, Aerospace Control and Application, 45, 3, pp. 53-58, (2019)
  • [6] FUJIMOTO S, HOOF H, MEGER D., Addressing function approximation error in Actor-Critic methods, Proceedings of International Conference on Machine Learning, pp. 1587-1596, (2018)
  • [7] FENG Chun, ZHANG Yiwei, HUANG Cheng, Et al., Deep reinforcement learning method for gait control of hiped rohot, Computer Integrated Manufacturing Systems, 27, 8, pp. 2341-2349, (2021)
  • [8] WANG Y, WANG L M, ZHAO Y H., Research on door opening operation of mohile rohotic arm based on reinforcement learning, Applied Sciences, 12, 10, (2022)
  • [9] CAI Ze, HU Yangguang, WEN Jingqian, Et al., AGV obstacle avoidance method based on deep reinforcement learning in complex dynamic environment, Computer Integrated Manufacturing Systems, 29, 1, pp. 236-245, (2023)
  • [10] XU J, HOU Z M, WANG W, Et al., Feedback deep deterministic policy gradient with fuzzy reward for robotic multiple peg-in-hole assembly tasks, Proceedings of the IEEE Transactions on Industrial Informatics, pp. 1658-1667, (2019)