Reinforcement learning for industrial process control: A case study in flatness control in steel industry

被引:48
作者
Deng, Jifei [1 ]
Sierla, Seppo [1 ]
Sun, Jie [2 ]
Vyatkin, Valeriy [1 ,3 ]
机构
[1] Aalto Univ, Sch Elect Engn, Dept Elect Engn & Automat, Espoo, Finland
[2] Northeastern Univ, State Key Lab Rolling & Automat, Shenyang, Peoples R China
[3] Lulea Univ Technol, Dept Comp Sci Elect & Space Engn, Lulea, Sweden
基金
中国国家自然科学基金;
关键词
Strip rolling; Process control; Reinforcement learning; Ensemble learning;
D O I
10.1016/j.compind.2022.103748
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Strip rolling is a typical manufacturing process, in which conventional control approaches are widely applied. Development of the control algorithms requires a mathematical expression of the process by means of the first principles or empirical models. However, it is difficult to upgrade the conventional control approaches in response to the ever-changing requirements and environmental conditions because domain knowledge of control engineering, mechanical engineering, and material science is required. Reinforcement learning is a machine learning method that can make the agent learn from interacting with the environment, thus avoiding the need for the above mentioned mathematical expression. This paper proposes a novel approach that combines ensemble learning with reinforcement learning methods for strip rolling control. Based on the proximal policy optimization (PPO), a multi-actor PPO is proposed. Each randomly initialized actor interacts with the environment in parallel, but only the experience from the actor that obtains the highest reward is used for updating the actors. Simulation results show that the proposed method outperforms the conventional control methods and the state-of-the-art reinforcement learning methods in terms of process capability and smoothness.
引用
收藏
页数:10
相关论文
共 38 条
[1]  
[Anonymous], 2016, INT C LEARN REPR ICL
[2]   A Deep Reinforcement Learning Approach to Improve the Learning Performance in Process Control [J].
Bao, Yaoyao ;
Zhu, Yuanming ;
Qian, Feng .
INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2021, 60 (15) :5504-5515
[3]   Optimization-based automatic flatness control in cold tandem rolling [J].
Bemporad, Alberto ;
Bernardini, Daniele ;
Cuzzola, Francesco Alessandro ;
Spinelli, Andrea .
JOURNAL OF PROCESS CONTROL, 2010, 20 (04) :396-407
[4]  
Box G.E., 2015, Time Series Analysis: Forecasting and Control
[5]   Model-Free Emergency Frequency Control Based on Reinforcement Learning [J].
Chen, Chunyu ;
Cui, Mingjian ;
Li, Fangxing ;
Yin, Shengfei ;
Wang, Xinan .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) :2336-2346
[6]   Application of neural networks for predicting hot-rolled strip crown [J].
Deng, Jifei ;
Sun, Jie ;
Peng, Wen ;
Hu, Yaohui ;
Zhang, Dianhua .
APPLIED SOFT COMPUTING, 2019, 78 :119-131
[7]   Deep-Reinforcement-Learning-Based Autonomous Voltage Control for Power Grid Operations [J].
Duan, Jiajun ;
Shi, Di ;
Diao, Ruisheng ;
Li, Haifeng ;
Wang, Zhiwei ;
Zhang, Bei ;
Bian, Desong ;
Yi, Zhehan .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2020, 35 (01) :814-817
[8]  
Fan HR, 2019, 2019 IEEE 4TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND BIG DATA ANALYSIS (ICCCBDA), P465, DOI [10.1109/icccbda.2019.8725683, 10.1109/ICCCBDA.2019.8725683]
[9]  
Fang Guo, 2020, Journal of Physics: Conference Series, V1486, DOI 10.1088/1742-6596/1486/7/072035
[10]  
Fujimoto S, 2018, PR MACH LEARN RES, V80