Reinforcement Learning-Based Optimal Fault-Tolerant Tracking Control of Industrial Processes

被引:6
作者
Wang, Limin [1 ,2 ]
Li, Xueyu [2 ]
Zhang, Ridong [3 ]
Gao, Furong [4 ]
机构
[1] Guangzhou Univ, Sch Mech & Elect Engn, Guangzhou 510006, Peoples R China
[2] Hainan Normal Univ, Sch Math & Stat, Haikou 571158, Peoples R China
[3] Hangzhou Dianzi Univ, Informat & Control Inst, Hangzhou 310018, Peoples R China
[4] Hong Kong Univ Sci & Technol, Dept Chem & Biol Engn, Hong Kong, Peoples R China
关键词
ADAPTIVE OPTIMAL-CONTROL; TIME LINEAR-SYSTEMS; NONLINEAR-SYSTEMS; DESIGN;
D O I
10.1021/acs.iecr.3c01789
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
For industrial processes with actuator faults, a reinforcement learning (RL)-based optimal fault-tolerant tracking control using a data-driven RL approach is proposed. First, a new state space model including tracking dynamics and state increment information is constructed. Subsequently, a performance index function is proposed based on the new model with actuator faults, followed by the derivation of the value function and Q-function. An RL algorithm is then designed to learn the optimal control law and optimize the performance index, resulting in the determination of the optimal controller gain. The proposed method holds promise, as it enables the expansion of the tolerable fault range of the system-a problem that has remained unresolved using reliable control techniques. It also achieves a good control effect even before the fault is eliminated, and the system performance is further improved through learning. The effectiveness and superiority of the proposed method are validated through a case study using a three-capacity water tank by comparing its performance with that of a model-based fault-tolerant tracking control method.
引用
收藏
页码:16014 / 16024
页数:11
相关论文
共 41 条
[1]  
Agrawal Y, 2020, INT CONF COMM SYST, P40, DOI [10.1109/CSNT48778.2020.9115758, 10.1109/CSNT.2020.08]
[2]   Fault-Tolerant Control of Degrading Systems with On-Policy Reinforcement Learning [J].
Ahmed, Ibrahim ;
Quinones-Grueiro, Marcos ;
Biswas, Gautam .
IFAC PAPERSONLINE, 2020, 53 (02) :13733-13738
[3]   Model-free Q-learning designs for linear discrete-time zero-sum games with application to H-infinity control [J].
Al-Tamimi, Asma ;
Lewis, Frank L. ;
Abu-Khalaf, Murad .
AUTOMATICA, 2007, 43 (03) :473-481
[4]   Robust fault tolerant tracking controller design for unknown inputs T-S models with unmeasurable premise variables [J].
Aouaouda, S. ;
Chadli, M. ;
Khadir, M. Tarek ;
Bouarar, T. .
JOURNAL OF PROCESS CONTROL, 2012, 22 (05) :861-872
[5]   Robust Model Predictive Control and Fault Handling of Batch Processes [J].
Aumi, Siam ;
Mhaskar, Prashant .
AICHE JOURNAL, 2011, 57 (07) :1796-1808
[6]   Decentralized fault-tolerant control system design for unstable processes [J].
Bao, J ;
Zhang, WZ ;
Lee, PL .
CHEMICAL ENGINEERING SCIENCE, 2003, 58 (22) :5045-5054
[7]   Iterative Learning Fault-Tolerant Control for Networked Batch Processes with Multirate Sampling and Quantization Effects [J].
Gao, Ming ;
Sheng, Li ;
Zhou, Donghua ;
Gao, Furong .
INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2017, 56 (09) :2515-2525
[8]   Learning-Based Adaptive Optimal Tracking Control of Strict-Feedback Nonlinear Systems [J].
Gao, Weinan ;
Jiang, Zhong-Ping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) :2614-2624
[9]   Output-feedback adaptive optimal control of interconnected systems based on robust adaptive dynamic programming [J].
Gao, Weinan ;
Jiang, Yu ;
Jiang, Zhong-Ping ;
Chai, Tianyou .
AUTOMATICA, 2016, 72 :37-45
[10]   Computational adaptive optimal control for continuous-time linear systems with completely unknown dynamics [J].
Jiang, Yu ;
Jiang, Zhong-Ping .
AUTOMATICA, 2012, 48 (10) :2699-2704