Optimal robust online tracking control for space manipulator in task space using off-policy reinforcement learning

被引:0
作者
Zhuang, Hongji [1 ]
Zhou, Hang [1 ]
Shen, Qiang [1 ]
Wu, Shufan [1 ,2 ]
Razoumny, Vladimir Yu. [2 ]
Razoumny, Yury N. [2 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai 200240, Peoples R China
[2] Peoples Friendship Univ Russia, RUDN Univ, Moscow 117198, Russia
基金
中国国家自然科学基金;
关键词
Task space; Reinforcement learning; Online tracking control; H-INFINITY CONTROL; ROBOT MANIPULATORS; NONLINEAR-SYSTEMS; TIME-SYSTEMS; APPROXIMATION; STATE;
D O I
10.1016/j.ast.2024.109446
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This study addresses the demands for adaptability, uncertainty management, and high performance in the control of space manipulators, and the inadequacies in achieving optimal control and handling external uncertainty in task space in previous research. Based on off-policy reinforcement learning, a model-free and time-efficient method for online robust tracking control in task space is devised. To address the complexity of dynamic equations in task space, a mixed-variable approach is adopted to transform the multivariable coupled time- varying problem into a single-variable problem. Subsequently, the optimal control policy is derived with the disturbance convergence, stability, and optimality of the control method being demonstrated. This marks the first instance of achieving robust optimal tracking control in task space for space manipulators. The efficacy and superiority of the presented algorithm are validated through simulation.
引用
收藏
页数:12
相关论文
共 50 条
[31]   Data-Driven Robust Control of Discrete-Time Uncertain Linear Systems via Off-Policy Reinforcement Learning [J].
Yang, Yongliang ;
Guo, Zhishan ;
Xiong, Haoyi ;
Ding, Da-Wei ;
Yin, Yixin ;
Wunsch, Donald C. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (12) :3735-3747
[32]   Task Space Control of Hydraulic Construction Machines Using Reinforcement Learning [J].
Lee, Hyung Joo ;
Brell-Cokcan, Sigrid .
HUMAN-FRIENDLY ROBOTICS 2023, HFR 2023, 2024, 29 :181-195
[33]   Synchronous optimal control method for nonlinear systems with saturating actuators and unknown dynamics using off-policy integral reinforcement learning [J].
Zhang, Zenglian ;
Song, Ruizhuo ;
Cao, Min .
NEUROCOMPUTING, 2019, 356 :162-169
[34]   Optimal model-free output synchronization of heterogeneous systems using off-policy reinforcement learning [J].
Modares, Hamidreza ;
Nageshrao, Subramanya P. ;
Lopes, Gabriel A. Delgado ;
Babuska, Robert ;
Lewis, Frank L. .
AUTOMATICA, 2016, 71 :334-341
[35]   H∞ Tracking Control of Completely Unknown Continuous-Time Systems via Off-Policy Reinforcement Learning [J].
Modares, Hamidreza ;
Lewis, Frank L. ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (10) :2550-2562
[36]   Model-free H control of Itô stochastic system via off-policy reinforcement learning [J].
Zhang, Weihai ;
Guo, Jing ;
Jiang, Xiushan .
AUTOMATICA, 2025, 174
[37]   Off-policy integral reinforcement learning-based optimal tracking control for a class of nonzero-sum game systems with unknown dynamics [J].
Zhao, Jin-Gang ;
Chen, Fang-Fang .
OPTIMAL CONTROL APPLICATIONS & METHODS, 2022, 43 (06) :1623-1644
[38]   Nonlinear tracking control on a robot manipulator in the task space with uncertain dynamics [J].
Soltanpour, M.R. ;
Fateh, M.M. ;
Ahmadi Fard, A.R. .
Journal of Applied Sciences, 2008, 8 (23) :4397-4403
[39]   Nearly Optimal Control for Mixed Zero-Sum Game Based on Off-Policy Integral Reinforcement Learning [J].
Song, Ruizhuo ;
Yang, Gaofu ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) :2793-2804
[40]   Optimal tracking control for discrete-time systems by model-free off-policy Q-learning approach [J].
Li, Jinna ;
Yuan, Decheng ;
Ding, Zhengtao .
2017 11TH ASIAN CONTROL CONFERENCE (ASCC), 2017, :7-12