Error analysis for physics-informed neural networks (PINNs) approximating Kolmogorov PDEs

被引:0
作者
Tim De Ryck
Siddhartha Mishra
机构
[1] Seminar for Applied Mathematics,
[2] ETH Zürich,undefined
来源
Advances in Computational Mathematics | 2022年 / 48卷
关键词
Physics-informed neural networks; Deep learning; Kolmogorov equations; 65M99;
D O I
暂无
中图分类号
学科分类号
摘要
Physics-informed neural networks approximate solutions of PDEs by minimizing pointwise residuals. We derive rigorous bounds on the error, incurred by PINNs in approximating the solutions of a large class of linear parabolic PDEs, namely Kolmogorov equations that include the heat equation and Black-Scholes equation of option pricing, as examples. We construct neural networks, whose PINN residual (generalization error) can be made as small as desired. We also prove that the total L2-error can be bounded by the generalization error, which in turn is bounded in terms of the training error, provided that a sufficient number of randomly chosen training (collocation) points is used. Moreover, we prove that the size of the PINNs and the number of training samples only grow polynomially with the underlying dimension, enabling PINNs to overcome the curse of dimensionality in this context. These results enable us to provide a comprehensive error analysis for PINNs in approximating Kolmogorov PDEs.
引用
收藏
相关论文
共 52 条
[1]  
Berner J(2020)Analysis of the generalization error: empirical risk minimization over deep artificial neural networks overcomes the curse of dimensionality in the numerical approximation of black–scholes partial differential equations SIAM J. Math. Data Sci. 2 631-657
[2]  
Grohs P(1995)Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems IEEE Trans. Neural Netw. 6 911-917
[3]  
Jentzen A(2021)On the approximation of functions by tanh neural networks Neural Netw. 143 732-750
[4]  
Chen T(2017)Deep learning-based numerical methods for high-dimensional parabolic partial differential equations and backward stochastic differential equations Commun. Math. Stat. 5 349-380
[5]  
Chen H(2020)Error bounds for approximations with deep ReLU neural networks in Ws,p norms Anal. Appl. 18 803-859
[6]  
De Ryck T(2021)Approximation rates for neural networks with encodable weights in smoothness spaces Neural Netw. 134 107-130
[7]  
Lanthaler S(2020)Extended physics-informed neural networks (XPINNs): A generalized space-time domain decomposition based deep learning framework for nonlinear partial differential equations Commun. Comput. Phys. 28 2002-2041
[8]  
Mishra S(2020)Conservative physics-informed neural networks on discrete domains for conservation laws: applications to forward and inverse problems Comput. Methods Appl. Mech. Eng. 365 113028-1049
[9]  
Han EW(2000)Neural-network methods for boundary value problems with irregular boundaries IEEE Trans. Neural Netw. 11 1041-1000
[10]  
Jentzen AJ(2000)Artificial neural networks for solving ordinary and partial differential equations IEEE Trans. Neural Netw. 9 987-228