Quantitative comparison of reinforcement learning and data-driven model predictive control for chemical and biological processes

被引:2
作者
Oh, Tae Hoon [1 ]
机构
[1] Kyoto Univ, Dept Chem Engn, Bldg A4, Katsura Campus, Nishikyo ku, Kyoto 6158510, Japan
关键词
Process control; Reinforcement learning; Model predictive control; Optimal control; System identification; NEURAL-NETWORKS; STABILITY;
D O I
10.1016/j.compchemeng.2023.108558
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
As manufacturing processes transition towards digitalization, data-driven process control is emerging as a key area of interest in future artificial intelligence technology. A crucial aspect in implementing data-driven process control is "What should we learn from the data?". In general, the data-driven control method can be categorized into two main approaches: Learning the model and learning the value. To assist in selecting the more suitable approach, this paper applies six different control methods, with three falling under each approach, to three distinct manufacturing process systems. The simulation results indicate that the model-learning approaches display higher data efficiency and exhibit lower variance in total cost. These methods prove to be particularly advantageous for addressing the regulation problems. Conversely, value-learning approaches show competitive potential in closed-loop identification and in managing economic cost problems. The remaining challenges associated with each technique are discussed, along with practical considerations for their implementation.
引用
收藏
页数:18
相关论文
共 52 条
  • [1] Allgower F., 1999, Advances in Control. Highlights of ECC'99, P391
  • [2] Bertsekas D. P., 2005, Dynamic Programming and Optimal Control
  • [3] Bock H.G., 1984, P IFAC WORLD C, P1603, DOI 10.1016/S1474-6670(17)61205-9
  • [4] On Recurrent Neural Networks for learning-based control: Recent results and ideas for future developments
    Bonassi, Fabio
    Farina, Marcello
    Xie, Jing
    Scattolini, Riccardo
    [J]. JOURNAL OF PROCESS CONTROL, 2022, 114 : 92 - 104
  • [5] Stability of discrete-time feed-forward neural networks in NARX configuration
    Bonassi, Fabio
    Farina, Marcello
    Scattolini, Riccardo
    [J]. IFAC PAPERSONLINE, 2021, 54 (07): : 547 - 552
  • [6] Brochu E, 2010, Arxiv, DOI arXiv:1012.2599
  • [7] Chen H., 1995, Proceedings of the Third European Control Conference. ECC 95, P3247
  • [8] Barsce JC, 2017, 2017 XLIII LATIN AMERICAN COMPUTER CONFERENCE (CLEI)
  • [9] Magnetic control of tokamak plasmas through deep reinforcement learning
    Degrave, Jonas
    Felici, Federico
    Buchli, Jonas
    Neunert, Michael
    Tracey, Brendan
    Carpanese, Francesco
    Ewalds, Timo
    Hafner, Roland
    Abdolmaleki, Abbas
    de las Casas, Diego
    Donner, Craig
    Fritz, Leslie
    Galperti, Cristian
    Huber, Andrea
    Keeling, James
    Tsimpoukelli, Maria
    Kay, Jackie
    Merle, Antoine
    Moret, Jean-Marc
    Noury, Seb
    Pesamosca, Federico
    Pfau, David
    Sauter, Olivier
    Sommariva, Cristian
    Coda, Stefano
    Duval, Basil
    Fasoli, Ambrogio
    Kohli, Pushmeet
    Kavukcuoglu, Koray
    Hassabis, Demis
    Riedmiller, Martin
    [J]. NATURE, 2022, 602 (7897) : 414 - +
  • [10] Fujimoto S, 2018, PR MACH LEARN RES, V80