Learning and meta-learning of stochastic advection-diffusion-reaction systems from sparse measurements

被引:30
作者
Chen, Xiaoli [1 ,2 ,3 ]
Duan, Jinqiao [4 ]
Emkarniadakis, George [3 ,5 ]
机构
[1] Huazhong Univ Sci & Technol, Ctr Math Sci, Wuhan 430074, Peoples R China
[2] Huazhong Univ Sci & Technol, Sch Math & Stat, Wuhan 430074, Peoples R China
[3] Brown Univ, Div Appl Math, Providence, RI 02912 USA
[4] IIT, Dept Appl Math, Chicago, IL 60616 USA
[5] Pacific Northwest Natl Lab, Richland, WA 99354 USA
关键词
Physics-informed neural networks; arbitrary polynomial chaos; multi-fidelity data; Karhunen– Loè ve expansion; uncertainty quantification; Bayesian optimisation; inverse problems; POLYNOMIAL CHAOS; APPROXIMATION; EQUATIONS;
D O I
10.1017/S0956792520000169
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Physics-informed neural networks (PINNs) were recently proposed in [18] as an alternative way to solve partial differential equations (PDEs). A neural network (NN) represents the solution, while a PDE-induced NN is coupled to the solution NN, and all differential operators are treated using automatic differentiation. Here, we first employ the standard PINN and a stochastic version, sPINN, to solve forward and inverse problems governed by a non-linear advection-diffusion-reaction (ADR) equation, assuming we have some sparse measurements of the concentration field at random or pre-selected locations. Subsequently, we attempt to optimise the hyper-parameters of sPINN by using the Bayesian optimisation method (meta-learning) and compare the results with the empirically selected hyper-parameters of sPINN. In particular, for the first part in solving the inverse deterministic ADR, we assume that we only have a few high-fidelity measurements, whereas the rest of the data is of lower fidelity. Hence, the PINN is trained using a composite multi-fidelity network, first introduced in [12], that learns the correlations between the multi-fidelity data and predicts the unknown values of diffusivity, transport velocity and two reaction constants as well as the concentration field. For the stochastic ADR, we employ a Karhunen-Loeve (KL) expansion to represent the stochastic diffusivity, and arbitrary polynomial chaos (aPC) to represent the stochastic solution. Correspondingly, we design multiple NNs to represent the mean of the solution and learn each aPC mode separately, whereas we employ a separate NN to represent the mean of diffusivity and another NN to learn all modes of the KL expansion. For the inverse problem, in addition to stochastic diffusivity and concentration fields, we also aim to obtain the (unknown) deterministic values of transport velocity and reaction constants. The available data correspond to 7spatial points for the diffusivity and 20 space-time points for the solution, both sampled 2000 times. We obtain good accuracy for the deterministic parameters of the order of 1-2% and excellent accuracy for the mean and variance of the stochastic fields, better than three digits of accuracy. In the second part, we consider the previous stochastic inverse problem, and we use Bayesian optimisation to find five hyper-parameters of sPINN, namely the width, depth and learning rate of two NNs for learning the modes. We obtain much deeper and wider optimal NNs compared to the manual tuning, leading to even better accuracy, i.e., errors less than 1% for the deterministic values, and about an order of magnitude less for the stochastic fields.
引用
收藏
页码:397 / 420
页数:24
相关论文
共 24 条
[1]  
BARAJAS-SOLANO D., 2018, ARXIV PREPRINT ARXIV
[2]   Approximate Bayesian model inversion for PDEs with heterogeneous and state-dependent coefficients [J].
Barajas-Solano, D. A. ;
Tartakovsky, A. M. .
JOURNAL OF COMPUTATIONAL PHYSICS, 2019, 395 :247-262
[3]  
Bergstra J., 2011, P 25 INT C NEUR INF, P24, DOI DOI 10.5555/2986459.2986743
[4]   Deep relaxation: partial differential equations for optimizing deep neural networks [J].
Chaudhari, Pratik ;
Oberman, Adam ;
Osher, Stanley ;
Soatto, Stefano ;
Carlier, Guillaume .
RESEARCH IN THE MATHEMATICAL SCIENCES, 2018, 5 :1-30
[5]  
Falkner S., 2018, ARXIV PREPRINT ARXIV
[6]  
Finn C, 2017, PR MACH LEARN RES, V70
[7]   Solving high-dimensional partial differential equations using deep learning [J].
Han, Jiequn ;
Jentzen, Arnulf ;
Weinan, E. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2018, 115 (34) :8505-8510
[8]   AMC: AutoML for Model Compression and Acceleration on Mobile Devices [J].
He, Yihui ;
Lin, Ji ;
Liu, Zhijian ;
Wang, Hanrui ;
Li, Li-Jia ;
Han, Song .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :815-832
[9]  
Li LS, 2018, J MACH LEARN RES, V18
[10]   COORDINATE DESCENT OPTIMIZATION FOR l1 MINIMIZATION WITH APPLICATION TO COMPRESSED SENSING; A GREEDY ALGORITHM [J].
Li, Yingying ;
Osher, Stanley .
INVERSE PROBLEMS AND IMAGING, 2009, 3 (03) :487-503