A STOCHASTIC NEWTON MCMC METHOD FOR LARGE-SCALE STATISTICAL INVERSE PROBLEMS WITH APPLICATION TO SEISMIC INVERSION

被引:291
作者
Martin, James [1 ]
Wilcox, Lucas C. [2 ]
Burstedde, Carsten
Ghattas, Omar [3 ,4 ]
机构
[1] Univ Texas Austin, Inst Computat Engn & Sci, Computat Sci Engn & Math Grad Program, Austin, TX 78712 USA
[2] USN, Postgrad Sch, Dept Appl Math, Monterey, CA 93943 USA
[3] Univ Texas Austin, Inst Computat Engn & Sci, Jackson Sch Geosci, Austin, TX 78712 USA
[4] Univ Texas Austin, Dept Mech Engn, Austin, TX 78712 USA
基金
美国国家科学基金会;
关键词
MCMC; Stochastic Newton; inverse problems; uncertainty quantification; Langevin dynamics; low-rank Hessian; MODEL-REDUCTION; POSTERIOR; CALIBRATION; ALGORITHMS; LANGEVIN;
D O I
10.1137/110845598
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We address the solution of large-scale statistical inverse problems in the framework of Bayesian inference. The Markov chain Monte Carlo (MCMC) method is the most popular approach for sampling the posterior probability distribution that describes the solution of the statistical inverse problem. MCMC methods face two central difficulties when applied to large-scale inverse problems: first, the forward models (typically in the form of partial differential equations) that map uncertain parameters to observable quantities make the evaluation of the probability density at any point in parameter space very expensive; and second, the high-dimensional parameter spaces that arise upon discretization of infinite-dimensional parameter fields make the exploration of the probability density function prohibitive. The challenge for MCMC methods is to construct proposal functions that simultaneously provide a good approximation of the target density while being inexpensive to manipulate. Here we present a so-called Stochastic Newton method in which MCMC is accelerated by constructing and sampling from a proposal density that builds a local Gaussian approximation based on local gradient and Hessian (of the log posterior) information. Thus, the method exploits tools (adjoint-based gradients and Hessians) that have been instrumental for fast (often mesh-independent) solution of deterministic inverse problems. Hessian manipulations (inverse, square root) are made tractable by a low-rank approximation that exploits the compact nature of the data misfit operator. This is analogous to a reduced model of the parameter-to-observable map. The method is applied to the Bayesian solution of an inverse medium problem governed by 1D seismic wave propagation. We compare the Stochastic Newton method with a reference black box MCMC method as well as a gradient-based Langevin MCMC method, and observe at least two orders of magnitude improvement in convergence for problems with up to 65 parameters. Numerical evidence suggests that a 1025 parameter problem converges at the same rate as the 65 parameter problem.
引用
收藏
页码:A1460 / A1487
页数:28
相关论文
共 47 条
[1]   Parallel algorithms for PDE-constrained optimization [J].
Akcelik, Volkan ;
Biros, George ;
Ghattas, Omar ;
Hill, Judith ;
Keyes, David ;
Waanders, Bart van Bloemen .
PARALLEL PROCESSING FOR SCIENTIFIC COMPUTING, 2006, :291-322
[2]   A General Purpose Sampling Algorithm for Continuous Distributions (the t-walk) [J].
Andres Christen, J. ;
Fox, Colin .
BAYESIAN ANALYSIS, 2010, 5 (02) :263-281
[3]  
[Anonymous], 1996, Cambridge Monographs on Applied and Computational Mathematics, DOI DOI 10.4310/CMS.2004.V2.N4.A7
[4]  
[Anonymous], 2003, ITERATIVE METHODS SP, DOI DOI 10.1137/1.9780898718003
[5]   Sampling the posterior: An approach to non-Gaussian data assimilation [J].
Apte, A. ;
Hairer, M. ;
Stuart, A. M. ;
Voss, J. .
PHYSICA D-NONLINEAR PHENOMENA, 2007, 230 (1-2) :50-64
[6]   Approximation errors and model reduction with an application in optical diffusion tomography [J].
Arridge, SR ;
Kaipio, JP ;
Kolehmainen, V ;
Schweiger, M ;
Somersalo, E ;
Tarvainen, T ;
Vauhkonen, M .
INVERSE PROBLEMS, 2006, 22 (01) :175-195
[7]   Parallel Lagrange-Newton-Krylov-Schur methods for PDE-constrained optimization. Part II: The Lagrange-Newton solver and its application to optimal control of steady viscous flows [J].
Biros, G ;
Ghattas, O .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 2005, 27 (02) :714-739
[8]   General methods for monitoring convergence of iterative simulations [J].
Brooks, SP ;
Gelman, A .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 1998, 7 (04) :434-455
[9]   Parametric Reduced-Order Models for Probabilistic Analysis of Unsteady Aerodynamic Applications [J].
Bui-Thanh, T. ;
Willcox, K. ;
Ghattas, O. .
AIAA JOURNAL, 2008, 46 (10) :2520-2529
[10]  
Bui-Thanh T., P SC12 GORD BE UNPUB