High performance binding affinity prediction with a Transformer-based surrogate model

被引:1
|
作者
Vasan, Archit [1 ]
Gokdemir, Ozan [1 ,2 ]
Brace, Alexander [1 ,2 ]
Ramanathan, Arvind [1 ,2 ]
Brettin, Thomas [1 ]
Stevens, Rick [1 ,2 ]
Vishwanath, Venkatram [1 ]
机构
[1] Argonne Natl Lab, Lemont, IL 60439 USA
[2] Univ Chicago, Chicago, IL 60637 USA
基金
美国国家卫生研究院;
关键词
drug discovery; virtual screening; docking surrogates; high performance computing; transformers; SMILES;
D O I
10.1109/IPDPSW63119.2024.00114
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In the current paradigm of drug discovery pipelines, identification of compounds that bind to a target with high affinity constitutes the first step. This is typically performed using resource-intensive experimental methods to screen vast chemical search spaces - a key bottleneck in the drug-discovery pipeline. To streamline this process, highly-scalable computational screening methods with acceptable fidelity are needed to screen larger portions of the chemical search space and identify promising candidates to he validated using experiments. Machine learning methods, namely, surrogate models have recently evolved into favorable alternatives to perform this computational screening. In this work, we present Simple SMILES Transformer (SST), an accurate and highly-scalable binding affinity prediction method that approximates the computationally-intensive molecular docking process using an encoder-only Transformer architecture. We benchmark our model against two baselines that feature fundamentally different approaches to docking surrogates: RegGO, a MORDRED fingerprint based multi-layer perceptron model, and Chemprop, a directed message-passing graph neural network. Unlike Chemprop and RegGO, our method operates solely on the SMILES representation of molecules without needing additional featurization, which leads to reduced preprocessing overhead, higher inference throughput and thus better scalability. We train SST in a distributed fashion on the Polaris supercomputer at the Argonne Leadership Computing Facility (ALCF). We then deploy it at an unprecedented scale for inference across 256 compute nodes of ALCF's Aurora supercomputer to screen 22 billion compounds in 40 minutes in search of hits with high binding affinity to oncoprotein RtcB ligase. SST predictions emphasize several molecular motifs that have previously been confirmed to interact with residues in their target binding pockets.
引用
收藏
页码:571 / 580
页数:10
相关论文
共 50 条
  • [31] An Explainable Transformer-Based Deep Learning Model for the Prediction of Incident Heart Failure
    Rao, Shishir
    Li, Yikuan
    Ramakrishnan, Rema
    Hassaine, Abdelaali
    Canoy, Dexter
    Cleland, John
    Lukasiewicz, Thomas
    Salimi-Khorshidi, Gholamreza
    Rahimi, Kazem
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (07) : 3362 - 3372
  • [32] MREDTA: A BERT and transformer-based molecular representation encoder for predicting drug-target binding affinity
    Sun, Xu
    Huang, Juanjuan
    Fang, Yabo
    Jin, Yixuan
    Wu, Jiageng
    Wang, Guoqing
    Jia, Jiwei
    FASEB JOURNAL, 2024, 38 (19):
  • [33] CityTransformer: A Transformer-Based Model for Contaminant Dispersion Prediction in a Realistic Urban Area
    Asahi, Yuuichi
    Onodera, Naoyuki
    Hasegawa, Yuta
    Shimokawabe, Takashi
    Shiba, Hayato
    Idomura, Yasuhiro
    BOUNDARY-LAYER METEOROLOGY, 2023, 186 (03) : 659 - 692
  • [34] Pedestrian Crossing Intention Prediction with Multi-Modal Transformer-Based Model
    Wang, Ting Wei
    Lai, Shang-Hong
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 1349 - 1356
  • [35] Transformer-Based Intelligent Prediction Model for Multimodal Multi-Objective Optimization
    Dang, Qianlong
    Zhang, Guanghui
    Wang, Ling
    Yu, Yang
    Yang, Shuai
    He, Xiaoyu
    IEEE COMPUTATIONAL INTELLIGENCE MAGAZINE, 2025, 20 (01) : 34 - 49
  • [36] Traffic Transformer: Transformer-based framework for temporal traffic accident prediction
    Al-Thani, Mansoor G.
    Sheng, Ziyu
    Cao, Yuting
    Yang, Yin
    AIMS MATHEMATICS, 2024, 9 (05): : 12610 - 12629
  • [37] Transformer and Graph Transformer-Based Prediction of Drug-Target Interactions
    Qian, Meiling
    Lu, Weizhong
    Zhang, Yu
    Liu, Junkai
    Wu, Hongjie
    Lu, Yaoyao
    Li, Haiou
    Fu, Qiming
    Shen, Jiyun
    Xiao, Yongbiao
    CURRENT BIOINFORMATICS, 2024, 19 (05) : 470 - 481
  • [38] MM-Transformer: A Transformer-Based Knowledge Graph Link Prediction Model That Fuses Multimodal Features
    Wang, Dongsheng
    Tang, Kangjie
    Zeng, Jun
    Pan, Yue
    Dai, Yun
    Li, Huige
    Han, Bin
    SYMMETRY-BASEL, 2024, 16 (08):
  • [39] A Transformer-Based Bridge Structural Response Prediction Framework
    Li, Ziqi
    Li, Dongsheng
    Sun, Tianshu
    SENSORS, 2022, 22 (08)
  • [40] Rethinking Transformer-based Set Prediction for Object Detection
    Sun, Zhiqing
    Cao, Shengcao
    Yang, Yiming
    Kitani, Kris
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 3591 - 3600