A Massively Parallel Infrastructure for Adaptive Multiscale Simulations: Modeling RAS Initiation Pathway for Cancer

被引:35
作者
Di Natale, Francesco [1 ]
Bhatia, Harsh [2 ]
Carpenter, Timothy S. [3 ]
Neale, Chris [4 ]
Kokkila-Schumacher, Sara [5 ]
Oppelstrup, Tomas [3 ]
Stanton, Liam [6 ]
Zhang, Xiaohua [3 ]
Sundram, Shiv [1 ]
Scogland, Thomas R. W. [2 ]
Dharuman, Gautham [3 ]
Surh, Michael P. [3 ]
Yang, Yue [3 ]
Misale, Claudia [5 ]
Schneidenbach, Lars [5 ]
Costa, Carlos [5 ]
Kim, Changhoan [5 ]
D'Amora, Bruce [5 ]
Gnanakaran, Sandrasegaram [4 ]
Nissley, Dwight, V [7 ]
Streitz, Fred [8 ]
Lightstone, Felice C. [3 ]
Bremer, Peer-Timo [2 ]
Glosli, James N. [3 ]
Ingolfsson, Helgi I. [3 ]
机构
[1] Lawrence Livermore Natl Lab, Applicat Simulat & Qual, Livermore, CA 94550 USA
[2] Lawrence Livermore Natl Lab, Ctr Appl Sci Comp, Livermore, CA 94550 USA
[3] Lawrence Livermore Natl Lab, Phys & Life Sci, Livermore, CA 94550 USA
[4] Los Alamos Natl Lab, Theoret Biol & Biophys, Los Alamos, NM 87545 USA
[5] IBM Corp, Thomas J Watson Res Ctr, Yorktown Hts, NY 10598 USA
[6] San Jose State Univ, Dept Math & Stat, San Jose, CA 95192 USA
[7] Frederick Natl Lab, Frederick, MD 21701 USA
[8] Lawrence Livermore Natl Lab, Livermore, CA 94550 USA
来源
PROCEEDINGS OF SC19: THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS | 2019年
基金
美国国家卫生研究院;
关键词
multiscale simulations; adaptive simulations; massively parallel; heterogenous architecture; machine learning; cancer research; MOLECULAR-DYNAMICS SIMULATIONS; COMPUTATIONAL LIPIDOMICS; MEMBRANE-PROTEIN; AGGREGATION; DENSITY; VERSION; SERIAL; FIELD;
D O I
10.1145/3295500.3356197
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Computational models can define the functional dynamics of complex systems in exceptional detail. However, many modeling studies face seemingly incommensurate requirements: to gain meaningful insights into some phenomena requires models with high resolution (microscopic) detail that must nevertheless evolve over large (macroscopic) length- and time-scales. Multiscale modeling has become increasingly important to bridge this gap. Executing complex multiscale models on current petascale computers with high levels of parallelism and heterogeneous architectures is challenging. Many distinct types of resources need to be simultaneously managed, such as GPUs and CPUs, memory size and latencies, communication bottlenecks, and filesystem bandwidth. In addition, robustness to failure of compute nodes, network, and filesystems is critical. We introduce a first-of-its-kind, massively parallel Multiscale Machine-Learned Modeling Infrastructure (MuMMI), which couples a macro scale model spanning micrometer length- and millisecond time-scales with a micro scale model employing high-fidelity molecular dynamics (MD) simulations. MuMMI is a cohesive and transferable infrastructure designed for scalability and efficient execution on heterogeneous resources. A central workflow manager simultaneously allocates GPIJs and CPUs while robustly handling failures in compute nodes, communication networks, and filesystems. A hierarchical scheduler controls GPU-accelerated MD simulations and in situ analysis. We present the various MuMMI components, including the macro model, GPU-accelerated MD, in situ analysis of MD data, machine learning selection module, a highly scalable hierarchical scheduler, and detail the central workflow manager that ties these modules together. In addition, we present performance data from our runs on Sierra, in which we validated MuMMI by investigating an experimentally intractable biological system: the dynamic interaction between RAS proteins and a plasma membrane. We used up to 4000 nodes of the Sierra supercomputer, concurrently utilizing over 16,000 GPUs and 176,000 CPU cores, and running up to 36,000 different tasks. This multiscale simulation includes about 120,000 MD simulations aggregating over 200 milliseconds, which is orders of magnitude greater than comparable studies.
引用
收藏
页数:16
相关论文
共 79 条
[1]   Gromacs: High performance molecular simulations through multi-level parallelism from laptops to supercomputers [J].
Abraham, Mark James ;
Murtola, Teemu ;
Schulz, Roland ;
Páll, Szilárd ;
Smith, Jeremy C. ;
Hess, Berk ;
Lindah, Erik .
SoftwareX, 2015, 1-2 :19-25
[2]   Flux: Overcoming Scheduling Challenges for Exascale Workflows [J].
Ahn, Dong H. ;
Bass, Ned ;
Chu, Albert ;
Garlick, Jim ;
Grondona, Mark ;
Herbein, Stephen ;
Koning, Joseph ;
Patki, Tapasya ;
Scogland, Thomas R. W. ;
Springmeyer, Becky ;
Taufer, Michela .
PROCEEDINGS OF WORKS 2018: 13TH IEEE/ACM WORKSHOP ON WORKFLOWS IN SUPPORT OF LARGE-SCALE SCIENCE (WORKS), 2018, :10-19
[3]  
Altintas I, 2004, 16TH INTERNATIONAL CONFERENCE ON SCIENTIFIC AND STATISTICAL DATABASE MANAGEMENT, PROCEEDINGS, P423
[4]   An Ensemble-Based Protocol for the Computational Prediction of Helix-Helix Interactions in G Protein-Coupled Receptors using Coarse-Grained Molecular Dynamics [J].
Altwaijry, Nojood A. ;
Baron, Michael ;
Wright, David W. ;
Coveney, Peter V. ;
Townsend-Nicholson, Andrea .
JOURNAL OF CHEMICAL THEORY AND COMPUTATION, 2017, 13 (05) :2254-2270
[6]  
[Anonymous], 2016, 100312013 IEEE, DOI DOI 10.1109/IEEESTD.2016.7582338
[7]  
[Anonymous], 1987, ComputerSimulation ofLiquids
[8]  
[Anonymous], 2019, JUN 2019 TOP500 SUP
[9]  
[Anonymous], 2009, Dakota, a multilevel parallel object-oriented framework for design optimization, parameter estimation, uncertainty quantification
[10]   Multiscale modeling of biomolecular systems: in serial and in parallel [J].
Ayton, Gary S. ;
Noid, Will G. ;
Voth, Gregory A. .
CURRENT OPINION IN STRUCTURAL BIOLOGY, 2007, 17 (02) :192-198