Optimized dead-zone inverse control using reinforcement learning and sliding-mode mechanism fora class of high-order nonlinear systems

被引:0
作者
Ma, Shuaihua [1 ]
Sun, Wenxia [1 ]
Wen, Guoxing [2 ]
机构
[1] Qilu Univ Technol, Shandong Acad Sci, Sch Math & Stat, Jinan 250353, Shandong, Peoples R China
[2] Shandong Univ Aeronaut, Coll Sci, Binzhou 256600, Shandong, Peoples R China
关键词
Optimal control; Dead-zone inverse; Reinforcement learning (RL); Sliding-mode mechanism; High-order nonlinear system; ROBUST ADAPTIVE-CONTROL; TRACKING CONTROL;
D O I
10.1016/j.ejcon.2024.101132
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
An optimized control method is developed fora class of high-order nonlinear dynamic systems having controller dead-zone phenomenon. Dead-zone refers to the controller with zero behavior within a certain range, so it will inevitably affect system performance. In order to make the optimized control eliminate the effect of dead zone, the adaptive dead-zone inverse and reinforcement learning (RL) techniques are combined. The main idea is to find the desired optimized control using RL as the input of dead-zone inverse function and then to design the adaptive algorithm to estimate the unknown parameters of dead-zone inverse function, so that the competent system control can be yielded from the dead-zone function. However, most existing RL algorithms are difficult to apply in the dead zone inverse methods because of the algorithm complexity. The proposed RL greatly simplifies the algorithm because it derives the training rules from the negative gradient of a simple positive function yielded by the partial derivative of Hamilton-Jacobi-Bellman (HJB) equation. Meanwhile, the proposed dead-zone inverse method also requires fewer adaptive parameters. Finally, the proposed control is attested through theoretical proofs and simulation examples.
引用
收藏
页数:8
相关论文
共 40 条
[1]   THE THEORY OF DYNAMIC PROGRAMMING [J].
BELLMAN, R .
BULLETIN OF THE AMERICAN MATHEMATICAL SOCIETY, 1954, 60 (06) :503-515
[2]   Reinforcement Learning and Adaptive Optimal Control for Continuous-Time Nonlinear Systems: A Value Iteration Approach [J].
Bian, Tao ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (07) :2781-2790
[3]   Decentralized prescribed performance adaptive tracking control for Markovian jump uncertain nonlinear systems with input saturation [J].
Chang, Ru ;
Fang, Yiming ;
Liu, Le ;
Li, Jianxiong .
INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2017, 31 (02) :255-274
[4]   Observer-Based Adaptive Backstepping Consensus Tracking Control for High-Order Nonlinear Semi-Strict-Feedback Multiagent Systems [J].
Chen, C. L. Philip ;
Wen, Guo-Xing ;
Liu, Yan-Jun ;
Liu, Zhi .
IEEE TRANSACTIONS ON CYBERNETICS, 2016, 46 (07) :1591-1601
[5]  
Cho HY, 1998, INT J ADAPT CONTROL, V12, P451, DOI 10.1002/(SICI)1099-1115(199808)12:5<451::AID-ACS504>3.0.CO
[6]  
2-R
[7]   Robust adaptive precision motion control of hydraulic actuators with valve dead-zone compensation [J].
Deng, Wenxiang ;
Yao, Jianyong ;
Ma, Dawei .
ISA TRANSACTIONS, 2017, 70 :269-278
[8]   Distributed adaptive containment control for high-order nonlinear multi-agent systems [J].
Gu, Jianzhong ;
Li, Wuquan ;
Yang, Hongyong .
INTERNATIONAL JOURNAL OF CONTROL, 2019, 92 (10) :2300-2311
[9]   Performance-Oriented Adaptive Robust Control of a Class of Nonlinear Systems Preceded by Unknown Dead Zone With Comparative Experimental Results [J].
Hu, Chuxiong ;
Yao, Bin ;
Wang, Qingfeng .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2013, 18 (01) :178-189
[10]  
Hull D. G., 2003, MECH ENG S