Comparison of Reinforcement Learning and Model Predictive Control for a Nonlinear Continuous Process

被引:0
作者
Rajpoot, Vikas [1 ]
Munusamy, Sudhakar [1 ]
Joshi, Tanuja [1 ]
Patil, Dinesh [1 ]
Pinnamaraju, Vivek [1 ]
机构
[1] ABB Corp Res, Bangalore, Karnataka, India
来源
IFAC PAPERSONLINE | 2024年 / 57卷
关键词
Reinforcement learning; Model predictive control; Nonlinear process; DDPG;
D O I
10.1016/j.ifacol.2024.05.052
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Model Predictive Control (MPC) has seen tremendous success in control of industrial processes due to its ability to effectively handle multi-input multi-output (MIMO) systems in the presence of process constraints. Effective control of nonlinear processes operated at wider operating regimes often requires either use of multiple linear models or a nonlinear model in the MPC framework. While theoretically this can result in improved performance compared to linear MPC, it suffers from additional complexities such as model switch scheduling, computational complexity, and convergence of solution to a local optimum. The Reinforcement Learning (RL) framework for control, which directly learns the control policy by interacting with the underlying process, is gaining growing interest, and is known to overcome the challenges faced by nonlinear MPC and achieve superior controller performance, with adequate exploration during training. In this work, we carry out a comparative analysis between RL and nonlinear MPC for a nonlinear chemical process - a Continuous Time Stirred Reactor (CSTR). Simulation studies reveal the superior performance of RL, attributed to its resolution of an infinite-horizon control problem, in contrast to MPC, which tackles finite-horizon optimization.
引用
收藏
页码:304 / 308
页数:5
相关论文
共 20 条
[1]   Continuous Control of Complex Chemical Reaction Network with Reinforcement Learning [J].
Alhazmi, Khalid ;
Sarathy, S. Mani .
2020 EUROPEAN CONTROL CONFERENCE (ECC 2020), 2020, :1066-1068
[2]  
Alhazmi Khalid, 2020, Diss
[3]   A Deep Reinforcement Learning Approach to Improve the Learning Performance in Process Control [J].
Bao, Yaoyao ;
Zhu, Yuanming ;
Qian, Feng .
INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2021, 60 (15) :5504-5515
[4]  
Edgar T.F., 2001, Optimization of Chemical Processes
[5]   Where Reinforcement Learning Meets Process Control: Review and Guidelines [J].
Faria, Ruan de Rezende ;
Olivier Capron, Bruno Didier ;
Secchi, Argimiro Resende ;
de Souza Jr, Mauricio B. .
PROCESSES, 2022, 10 (11)
[6]   Reinforcement learning in feedback control Challenges and benchmarks from technical process control [J].
Hafner, Roland ;
Riedmiller, Martin .
MACHINE LEARNING, 2011, 84 (1-2) :137-169
[7]   PROCESS-CONTROL VIA ARTIFICIAL NEURAL NETWORKS AND REINFORCEMENT LEARNING [J].
HOSKINS, JC ;
HIMMELBLAU, DM .
COMPUTERS & CHEMICAL ENGINEERING, 1992, 16 (04) :241-251
[8]   TASAC: A twin-actor reinforcement learning framework with a stochastic with an to batch control [J].
Joshi, Tanuja ;
Kodamana, Hariprasad ;
Kandath, Harikumar ;
Kaisare, Niket .
CONTROL ENGINEERING PRACTICE, 2023, 134
[9]   Twin actor twin delayed deep deterministic policy gradient (TATD3) learning for batch process control [J].
Joshi, Tanuja ;
Makker, Shikhar ;
Kodamana, Hariprasad ;
Kandath, Harikumar .
COMPUTERS & CHEMICAL ENGINEERING, 2021, 155
[10]  
Lillicrap T.P., 2015, arXiv