Reinforcement learning-based estimation for spatio-temporal systems

被引:1
|
作者
Mowlavi, Saviz [1 ]
Benosman, Mouhacine [1 ]
机构
[1] Mitsubishi Elect Res Labs, Cambridge, MA 02139 USA
来源
SCIENTIFIC REPORTS | 2024年 / 14卷 / 01期
关键词
Estimation; Filtering; Partial differential equations; Model reduction; Reinforcement learning; MODEL-REDUCTION; FLUID-FLOWS;
D O I
10.1038/s41598-024-72055-1
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
State estimators such as Kalman filters compute an estimate of the instantaneous state of a dynamical system from sparse sensor measurements. For spatio-temporal systems, whose dynamics are governed by partial differential equations (PDEs), state estimators are typically designed based on a reduced-order model (ROM) that projects the original high-dimensional PDE onto a computationally tractable low-dimensional space. However, ROMs are prone to large errors, which negatively affects the performance of the estimator. Here, we introduce the reinforcement learning reduced-order estimator (RL-ROE), a ROM-based estimator in which the correction term that takes in the measurements is given by a nonlinear policy trained through reinforcement learning. The nonlinearity of the policy enables the RL-ROE to compensate efficiently for errors of the ROM, while still taking advantage of the imperfect knowledge of the dynamics. Using examples involving the Burgers and Navier-Stokes equations with parametric uncertainties, we show that in the limit of very few sensors, the trained RL-ROE outperforms a Kalman filter designed using the same ROM and yields accurate instantaneous estimates of high-dimensional states corresponding to unknown initial conditions and physical parameter values. The RL-ROE opens the door to lightweight real-time sensing of systems governed by parametric PDEs.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Reinforcement learning-based estimation for spatio-temporal systems (vol 14, 22464, 2024)
    Mowlavi, Saviz
    Benosman, Mouhacine
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [2] Learning-based spatio-temporal vehicle tracking and indexing for transportation multimedia database systems
    Chen, SC
    Shyu, ML
    Peeta, S
    Zhang, CC
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2003, 4 (03) : 154 - 167
  • [3] A survey on deep learning-based spatio-temporal action detection
    Wang, Peng
    Zeng, Fanwei
    Qian, Yuntao
    INTERNATIONAL JOURNAL OF WAVELETS MULTIRESOLUTION AND INFORMATION PROCESSING, 2024, 22 (04)
  • [4] A task allocation algorithm based on reinforcement learning in spatio-temporal crowdsourcing
    Bingxu Zhao
    Hongbin Dong
    Yingjie Wang
    Tingwei Pan
    Applied Intelligence, 2023, 53 : 13452 - 13469
  • [5] A task allocation algorithm based on reinforcement learning in spatio-temporal crowdsourcing
    Zhao, Bingxu
    Dong, Hongbin
    Wang, Yingjie
    Pan, Tingwei
    APPLIED INTELLIGENCE, 2023, 53 (11) : 13452 - 13469
  • [6] Parallel Computing of Spatio-Temporal Model Based on Deep Reinforcement Learning
    Lv, Zhiqiang
    Li, Jianbo
    Xu, Zhihao
    Wang, Yue
    Li, Haoran
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2021, PT I, 2021, 12937 : 391 - 403
  • [7] An Attentive Spatio-Temporal Learning-Based Network for Cardiovascular Disease Diagnosis
    Jyotishi, Debasish
    Dandapat, Samarendra
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (08): : 4661 - 4671
  • [8] Learning-Based Distributed Spatio-Temporal k Nearest Neighbors Join
    Li, Ruiyuan
    Li, Jiajun
    Zhou, Minxin
    Wang, Rubin
    He, Huajun
    Chen, Chao
    Bao, Jie
    Zheng, Yu
    IEEE TRANSACTIONS ON BIG DATA, 2025, 11 (02) : 861 - 878
  • [9] Quantum Reinforcement Learning for Spatio-Temporal Prioritization in Metaverse
    Park, Soohyun
    Baek, Hankyul
    Kim, Joongheon
    IEEE ACCESS, 2024, 12 : 54732 - 54744
  • [10] Estimating spatio-temporal fields through reinforcement learning
    Padrao, Paulo
    Fuentes, Jose
    Bobadilla, Leonardo
    Smith, Ryan N.
    FRONTIERS IN ROBOTICS AND AI, 2022, 9