Local Control is All You Need: Decentralizing and Coordinating Reinforcement Learning for Large-Scale Process Control

被引:0
作者
Bougie, Nicolas [1 ]
Onishi, Takashi [1 ,2 ]
Tsuruoka, Yoshimasa [1 ,3 ]
机构
[1] Natl Inst Adv Ind Sci & Technol, INEC AIST Cooperat Res Lab, Tokyo, Japan
[2] NEC Corp Ltd, Data Sci Res Labs, Kawasaki, Kanagawa, Japan
[3] Univ Tokyo, Dept Informat & Commun Engn, Tokyo, Japan
来源
2022 61ST ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS (SICE) | 2022年
关键词
deep reinforcement learning; large-scale reinforcement learning; chemical process control;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep reinforcement learning (RL) approaches are an appealing alternative to conventional controllers in process industries as such methods are inherently flexible and have generalization abilities to unseen situations. Namely, they alleviate the need for constant parameter tuning, tedious design of control laws, and re-identification procedures in the event of performance degradation. However, it remains challenging to apply RL to real-world process tasks, which commonly feature large state-action spaces and complex dynamics. Such tasks may be difficult to solve due to computational complexity and sample insufficiency. To tackle these limitations, we present a sample-efficient RL approach for large-scale control that expresses the global policy as a collection of local policies. Every local policy receives local observations and is responsible for controlling a different region of the environment. In order to enable coordination among local policies, we present a mechanism based on action sharing and message passing. The model is evaluated on a set of robotic tasks and a large-scale vinyl acetate monomer (VAM) plant. The experiments demonstrate that the proposed model exhibits significant improvements over baselines in terms of mean scores and sample efficiency.
引用
收藏
页码:468 / 474
页数:7
相关论文
共 18 条
  • [1] Cui YD, 2018, IEEE INT CON AUTO SC, P304, DOI 10.1109/COASE.2018.8560593
  • [2] Duan Y, 2016, PR MACH LEARN RES, V48
  • [3] Haarnoja T, 2019, Arxiv, DOI arXiv:1812.05905
  • [4] Huang W., 2020, INT C MACHINE LEARNI, P4455
  • [5] Jaques N, 2019, PR MACH LEARN RES, V97
  • [6] Kingma DP, 2014, ADV NEUR IN, V27
  • [7] Kubosawa S., ANN C SOC INSTRUMENT, P21
  • [8] Kubosawa S, 2019, Arxiv, DOI arXiv:1903.02183
  • [9] Computing operation procedures for chemical plants using whole-plant simulation models
    Kubosawa, Shumpei
    Onishi, Takashi
    Tsuruoka, Yoshimasa
    [J]. CONTROL ENGINEERING PRACTICE, 2021, 114
  • [10] Vinyl Acetate Monomer (VAM) Plant Model: A New Benchmark Problem for Control and Operation Study
    Machida, Yuta
    Ootakara, Shigeki
    Seki, Hiroya
    Hashimoto, Yoshihiro
    Kano, Manabu
    Miyake, Yasuhiro
    Anzai, Naoto
    Sawai, Masavoshi
    Katsuno, Takashi
    Omata, Toshiaki
    [J]. IFAC PAPERSONLINE, 2016, 49 (07): : 533 - 538