An autoencoder compression approach for accelerating large-scale inverse problems

被引:0
作者
Wittmer, Jonathan [1 ]
Badger, Jacob [1 ]
Sundar, Hari [2 ]
Bui-Thanh, Tan [3 ]
机构
[1] Univ Texas Austin, Oden Inst Computat Sci & Engn, Austin, TX 78712 USA
[2] Univ Utah, Dept Comp Sci, Salt Lake City, UT 84112 USA
[3] Univ Texas Austin, Dept Aerosp Engn & Engn Mech, Oden Inst Computat Sci & Engn, Austin, TX 78712 USA
基金
美国国家科学基金会;
关键词
machine learning; autoencoder; inverse problems; high performance computing; BIG DATA; ADJOINT;
D O I
10.1088/1361-6420/acfbe1
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Partial differential equation (PDE)-constrained inverse problems are some of the most challenging and computationally demanding problems in computational science today. Fine meshes required to accurately compute the PDE solution introduce an enormous number of parameters and require large-scale computing resources such as more processors and more memory to solve such systems in a reasonable time. For inverse problems constrained by time-dependent PDEs, the adjoint method often employed to compute gradients and higher order derivatives efficiently requires solving a time-reversed, so-called adjoint PDE that depends on the forward PDE solution at each timestep. This necessitates the storage of a high-dimensional forward solution vector at every timestep. Such a procedure quickly exhausts the available memory resources. Several approaches that trade additional computation for reduced memory footprint have been proposed to mitigate the memory bottleneck, including checkpointing and compression strategies. In this work, we propose a close-to-ideal scalable compression approach using autoencoders to eliminate the need for checkpointing and substantial memory storage, thereby reducing the time-to-solution and memory requirements. We compare our approach with checkpointing and an off-the-shelf compression approach on an earth-scale ill-posed seismic inverse problem. The results verify the expected close-to-ideal speedup for the gradient and Hessian-vector product using the proposed autoencoder compression approach. To highlight the usefulness of the proposed approach, we combine the autoencoder compression with the data-informed active subspace (DIAS) prior showing how the DIAS method can be affordably extended to large-scale problems without the need for checkpointing and large memory.
引用
收藏
页数:26
相关论文
共 70 条
[1]  
Abu Alsheikh M., 2014, PROC 17 ACM INT C MO, ppp 307
[2]   Trading Computation for Communication: A Taxonomy of Data Recomputation Techniques [J].
Akturk, Ismail ;
Karpuzcu, Ulya R. .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2021, 9 (01) :496-506
[3]  
[Anonymous], 2016, Matrix Analysis for Statistics
[4]   NEURAL NETWORKS AND PRINCIPAL COMPONENT ANALYSIS - LEARNING FROM EXAMPLES WITHOUT LOCAL MINIMA [J].
BALDI, P ;
HORNIK, K .
NEURAL NETWORKS, 1989, 2 (01) :53-58
[5]  
Bharadwaj K K., 2021, Computational Intelligence Methods in Covid-19: Surveillance, Prevention, Prediction and Diagnosis, ppp 311
[6]   Preface to Special Topic: Building the Bridge to the Exascale-Applications and Opportunities for Plasma Physics [J].
Bhattacharjee, Amitava ;
Wells, Jack .
PHYSICS OF PLASMAS, 2021, 28 (09)
[7]   Kohn-Sham density functional theory: Predicting and understanding chemistry [J].
Bickelhaupt, FM ;
Baerends, EJ .
REVIEWS IN COMPUTATIONAL CHEMISTRY, VOL 15, 2000, 15 :1-86
[8]   Real-time insect tracking and monitoring with computer vision and deep learning [J].
Bjerge, Kim ;
Mann, Hjalte M. R. ;
Hoye, Toke Thomas .
REMOTE SENSING IN ECOLOGY AND CONSERVATION, 2022, 8 (03) :315-327
[9]   AUDIO DYNAMIC RANGE COMPRESSION FOR MINIMUM PERCEIVED DISTORTION [J].
BLESSER, BA .
IEEE TRANSACTIONS ON AUDIO AND ELECTROACOUSTICS, 1969, AU17 (01) :22-&
[10]   Wavefield compression for adjoint methods in full-waveform inversion [J].
Boehm, Christian ;
Hanzich, Mauricio ;
de la Puente, Josep ;
Fichtner, Andreas .
GEOPHYSICS, 2016, 81 (06) :R385-R397