End-to-end reinforcement learning of Koopman models for economic nonlinear model predictive control

被引:0
作者
Mayfrank, Daniel [1 ,4 ]
Mitsos, Alexander [1 ,2 ,3 ]
Dahmen, Manuel [1 ]
机构
[1] Forschungszentrum Julich, Inst Climate & Energy Syst Energy Syst Engn ICE 1, D-52425 Julich, Germany
[2] Rhein Westfal TH Aachen, Proc Syst Engn AVTSVT, D-52074 Aachen, Germany
[3] JARA ENERGY, D-52425 Julich, Germany
[4] Rhein Westfal TH Aachen, D-52062 Aachen, Germany
关键词
Economic model predictive control; Koopman; Reinforcement learning; End-to-end learning; OPERATOR; SYSTEMS;
D O I
10.1016/j.compchemeng.2024.108824
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
(Economic) nonlinear model predictive control ((e)NMPC) requires dynamic models that are sufficiently accurate and computationally tractable. Data-driven surrogate models for mechanistic models can reduce the computational burden of (e)NMPC; however, such models are typically trained by system identification for maximum prediction accuracy on simulation samples and perform suboptimally in (e)NMPC. We present a method for end-to-end reinforcement learning of Koopman surrogate models for optimal performance as part of (e)NMPC. We apply our method to two applications derived from an established nonlinear continuous stirred- tank reactor model. The controller performance is compared to that of (e)NMPCs utilizing models trained using system identification, and model-free neural network controllers trained using reinforcement learning. We show that the end-to-end trained models outperform those trained using system identification in (e)NMPC, and that, in contrast to the neural network controllers, the (e)NMPC controllers can react to changes in the control setting without retraining.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Reinforcement Learning-Based End-to-End Parking for Automatic Parking System
    Zhang, Peizhi
    Xiong, Lu
    Yu, Zhuoping
    Fang, Peiyuan
    Yan, Senwei
    Yao, Jie
    Zhou, Yi
    [J]. SENSORS, 2019, 19 (18)
  • [32] Latency Equalization Policy of End-to-End Network Slicing Based on Reinforcement Learning
    Bai, Haonan
    Zhang, Yong
    Zhang, Zhenyu
    Yuan, Siyu
    [J]. IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2023, 20 (01): : 88 - 103
  • [33] Incremental End-to-End Learning for Lateral Control in Autonomous Driving
    Kwon, Jaerock
    Khalil, Aws
    Kim, Donghyun
    Nam, Haewoon
    [J]. IEEE ACCESS, 2022, 10 : 33771 - 33786
  • [34] Efficient economic model predictive control of water treatment process with learning-based Koopman operator
    Han, Minghao
    Yao, Jingshi
    Law, Adrian Wing-Keung
    Yin, Xunyuan
    [J]. CONTROL ENGINEERING PRACTICE, 2024, 149
  • [35] An End-to-End Path Planner Combining Potential Field Method With Deep Reinforcement Learning
    Wang, Yixuan
    Shen, Bin
    Nan, Zhuojiang
    Tao, Wei
    [J]. IEEE SENSORS JOURNAL, 2024, 24 (16) : 26584 - 26591
  • [36] Characterizing and Optimizing the End-to-End Performance of Multi-Agent Reinforcement Learning Systems
    Gogineni, Kailash
    Mei, Yongsheng
    Gogineni, Karthikeya
    Wei, Peng
    Lan, Tian
    Venkataramani, Guru
    [J]. 2024 IEEE INTERNATIONAL SYMPOSIUM ON WORKLOAD CHARACTERIZATION, IISWC 2024, 2024, : 224 - 235
  • [37] A Vision-Based End-to-End Reinforcement Learning Framework for Drone Target Tracking
    Zhao, Xun
    Huang, Xinjian
    Cheng, Jianheng
    Xia, Zhendong
    Tu, Zhiheng
    [J]. DRONES, 2024, 8 (11)
  • [38] Model-Based End-to-End Learning for WDM Systems With Transceiver Hardware Impairments
    Song, Jinxiang
    Haeger, Christian
    Schroeder, Jochen
    Amat, Alexandre Graell, I
    Wymeersch, Henk
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN QUANTUM ELECTRONICS, 2022, 28 (04)
  • [39] Nonlinear Dual-Mode Model Predictive Control using Koopman Eigenfunctions
    Krolicki, Alexander
    Tellez-Castro, Duvan
    Vaidya, Umesh
    [J]. 2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 3074 - 3079
  • [40] Autonomous Vehicle Control: End-to-End Learning in Simulated Urban Environments
    Haavaldsen, Hege
    Aasbo, Max
    Lindseth, Frank
    [J]. NORDIC ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2019, 1056 : 40 - 51