High performance binding affinity prediction with a Transformer-based surrogate model

被引:1
|
作者
Vasan, Archit [1 ]
Gokdemir, Ozan [1 ,2 ]
Brace, Alexander [1 ,2 ]
Ramanathan, Arvind [1 ,2 ]
Brettin, Thomas [1 ]
Stevens, Rick [1 ,2 ]
Vishwanath, Venkatram [1 ]
机构
[1] Argonne Natl Lab, Lemont, IL 60439 USA
[2] Univ Chicago, Chicago, IL 60637 USA
基金
美国国家卫生研究院;
关键词
drug discovery; virtual screening; docking surrogates; high performance computing; transformers; SMILES;
D O I
10.1109/IPDPSW63119.2024.00114
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In the current paradigm of drug discovery pipelines, identification of compounds that bind to a target with high affinity constitutes the first step. This is typically performed using resource-intensive experimental methods to screen vast chemical search spaces - a key bottleneck in the drug-discovery pipeline. To streamline this process, highly-scalable computational screening methods with acceptable fidelity are needed to screen larger portions of the chemical search space and identify promising candidates to he validated using experiments. Machine learning methods, namely, surrogate models have recently evolved into favorable alternatives to perform this computational screening. In this work, we present Simple SMILES Transformer (SST), an accurate and highly-scalable binding affinity prediction method that approximates the computationally-intensive molecular docking process using an encoder-only Transformer architecture. We benchmark our model against two baselines that feature fundamentally different approaches to docking surrogates: RegGO, a MORDRED fingerprint based multi-layer perceptron model, and Chemprop, a directed message-passing graph neural network. Unlike Chemprop and RegGO, our method operates solely on the SMILES representation of molecules without needing additional featurization, which leads to reduced preprocessing overhead, higher inference throughput and thus better scalability. We train SST in a distributed fashion on the Polaris supercomputer at the Argonne Leadership Computing Facility (ALCF). We then deploy it at an unprecedented scale for inference across 256 compute nodes of ALCF's Aurora supercomputer to screen 22 billion compounds in 40 minutes in search of hits with high binding affinity to oncoprotein RtcB ligase. SST predictions emphasize several molecular motifs that have previously been confirmed to interact with residues in their target binding pockets.
引用
收藏
页码:571 / 580
页数:10
相关论文
共 50 条
  • [41] An efficient transformer-based surrogate model with end-to-end training strategies for automatic history matching
    Zhang, Jinding
    Kang, Jinzheng
    Zhang, Kai
    Zhang, Liming
    Liu, Piyang
    Liu, Xingyu
    Sun, Weijia
    Wang, Guangyao
    GEOENERGY SCIENCE AND ENGINEERING, 2024, 240
  • [42] Transformer-based attention network for stock movement prediction
    Zhang, Qiuyue
    Qin, Chao
    Zhang, Yunfeng
    Bao, Fangxun
    Zhang, Caiming
    Liu, Peide
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 202
  • [43] TransCFD: A transformer-based decoder for flow field prediction
    Jiang, Jundou
    Li, Guanxiong
    Jiang, Yi
    Zhang, Laiping
    Deng, Xiaogang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 123
  • [44] Deep Transformer-Based Asset Price and Direction Prediction
    Gezici, Abdul Haluk Batur
    Sefer, Emre
    IEEE ACCESS, 2024, 12 : 24164 - 24178
  • [45] Transformer-based Architecture for Empathy Prediction and Emotion Classification
    Vasava, Himil
    Uikey, Pramegh
    Wasnik, Gaurav
    Sharma, Raksha
    PROCEEDINGS OF THE 12TH WORKSHOP ON COMPUTATIONAL APPROACHES TO SUBJECTIVITY, SENTIMENT & SOCIAL MEDIA ANALYSIS, 2022, : 261 - 264
  • [46] HTTNet: hybrid transformer-based approaches for trajectory prediction
    Ge, Xianlei
    Shen, Xiaobo
    Zhou, Xuanxin
    Li, Xiaoyan
    BULLETIN OF THE POLISH ACADEMY OF SCIENCES-TECHNICAL SCIENCES, 2024, 72 (05)
  • [47] Transformer-Based Prediction of Hospital Readmissions for Diabetes Patients
    Garcia-Mosquera, Jorge
    Villa-Monedero, Maria
    Gil-Martin, Manuel
    San-Segundo, Ruben
    ELECTRONICS, 2025, 14 (01):
  • [48] SST: A Simplified Swin Transformer-based Model for Taxi Destination Prediction based on Existing Trajectory
    Wang, Zepu
    Sun, Yifei
    Lei, Zhiyu
    Zhu, Xincheng
    Sun, Peng
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 1404 - 1409
  • [49] Improving transformer-based acoustic model performance using sequence discriminative training
    Lee, Chae-Won
    Chang, Joon-Hyuk
    JOURNAL OF THE ACOUSTICAL SOCIETY OF KOREA, 2022, 41 (03): : 335 - 341
  • [50] The Application of a BiGRU Model with Transformer-Based Error Correction in Deformation Prediction for Bridge SHM
    Wang, Xu
    Xie, Guilin
    Zhang, Youjia
    Liu, Haiming
    Zhou, Lei
    Liu, Wentao
    Gao, Yang
    BUILDINGS, 2025, 15 (04)