Multi-fidelity surrogate modeling using long short-term memory networks

被引:43
作者
Conti, Paolo [1 ]
Guo, Mengwu [2 ]
Manzoni, Andrea [3 ]
Hesthaven, Jan S. [4 ]
机构
[1] Politecn Milan, Dept Civil Engn, Milan, Italy
[2] Univ Twente, Dept Appl Math, Enschede, Netherlands
[3] Politecn Milan, Dept Math, MOX, Milan, Italy
[4] Ecole Polytech Fed Lausanne, Inst Math, Lausanne, Switzerland
关键词
Machine learning; Multi-fidelity regression; LSTM network; Parametrized PDE; Time-dependent problem; APPROXIMATION;
D O I
10.1016/j.cma.2022.115811
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
When evaluating quantities of interest that depend on the solutions to differential equations, we inevitably face the trade-off between accuracy and efficiency. Especially for parametrized, time-dependent problems in engineering computations, it is often the case that acceptable computational budgets limit the availability of high-fidelity, accurate simulation data. Multi-fidelity surrogate modeling has emerged as an effective strategy to overcome this difficulty. Its key idea is to leverage many low-fidelity simulation data, less accurate but much faster to compute, to improve the approximations with limited high-fidelity data. In this work, we introduce a novel data-driven framework of multi-fidelity surrogate modeling for parametrized, time-dependent problems using long short-term memory (LSTM) networks, to enhance output predictions both for unseen parameter values and forward in time simultaneously - a task known to be particularly challenging for data-driven models. We demonstrate the wide applicability of the proposed approaches in a variety of engineering problems with high-and low-fidelity data generated through fine versus coarse meshes, small versus large time steps, or finite element full order versus deep learning reduced-order models. Numerical results show that the proposed multi-fidelity LSTM networks not only improve single-fidelity regression significantly, but also outperform the multi-fidelity models based on feed-forward neural networks.(c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:22
相关论文
共 49 条
[1]   Kernels for Vector-Valued Functions: A Review [J].
Alvarez, Mauricio A. ;
Rosasco, Lorenzo ;
Lawrence, Neil D. .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2012, 4 (03) :195-266
[2]  
[Anonymous], 2017, Deep Neural Networks as Gaussian Processes
[3]  
Baker N., 2019, Tech. Rep.
[4]   Machine Learning Approximation Algorithms for High-Dimensional Fully Nonlinear Partial Differential Equations and Second-order Backward Stochastic Differential Equations [J].
Beck, Christian ;
Weinan, E. ;
Jentzen, Arnulf .
JOURNAL OF NONLINEAR SCIENCE, 2019, 29 (04) :1563-1619
[5]  
Bergstra J., 2013, Proc. Mach. Learn. Res., V28
[6]  
Bergstra J., 2013, HYPEROPT DISTRIBUTED
[7]  
Bergstra J., 2011, Advances in Neural Information Processing Systems, V24, P2546
[8]   Data-driven discovery of coordinates and governing equations [J].
Champion, Kathleen ;
Lusch, Bethany ;
Kutz, J. Nathan ;
Brunton, Steven L. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (45) :22445-22451
[9]  
Cutajar K, 2019, Arxiv, DOI [arXiv:1903.07320, 10.48550/arXiv.1903.07320]
[10]  
Demo N, 2023, Arxiv, DOI arXiv:2110.13530