A center-anchored adaptive hierarchical graph neural network with application in structure-aware recognition of enzyme catalytic specificity

被引:0
作者
Li, Yi [1 ]
Liu, Yahui [1 ]
Hou, Jiangchun [2 ]
Liu, Xiaohui [3 ]
机构
[1] Dali Univ, Coll Math & Comp Sci, Dali, Peoples R China
[2] Yunnan Univ, Affiliated Hosp, Kunming, Peoples R China
[3] Brunel Univ London, Dept Comp Sci, Uxbridge, England
关键词
Graph Neural Networks; Adaptive hierarchical strategy; Two-step sampling method; Central anchoring; Enzyme modeling; PROTEIN; SYNTHASES; RESIDUE;
D O I
10.1016/j.neucom.2024.129155
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Non-Euclidean modeling of biomolecules such as enzymes has been advanced by utilizing graph neural networks (GNNs). While existing whole-molecule modeling approaches offer significant advantages, they present two primary challenges: optimizing long-range interactions within the enzyme molecule and ensuring model interpretability. Furthermore, current GNN-based enzyme modeling methods pay insufficient attention to the hierarchical structure of biological functions. To address these challenges, we propose a novel Center-Anchored Adaptive Hierarchical GNN (CAAH-GNN) model that integrates a center-anchored method with an adaptive hierarchical strategy for structure-aware recognition of enzyme catalytic specificity. Our model features three key characteristics: (1) It autonomously identifies and integrates local key residues within enzyme molecules while embedding global graph information into each node's features, significantly enhancing prediction accuracy. (2) The spherical spatial radius is adaptively learned using a two-step hierarchically sampling method. Compared to full-molecule models, this approach reduces spatial complexity more than 80 %. (3) Our model exhibits strong generalization performance and interpretability, as it is not tailored to specific enzyme families and does not remove any nodes during training. Extensive experiments based on two enzyme datasets and various graph convolution kernels demonstrate that CAAH-GNN outperforms baseline models and other hierarchical strategy models in prediction accuracy. In summary, our research not only introduces a novel GNNbased enzyme modeling approach but also provides a new perspective and a tool for enzyme engineering.
引用
收藏
页数:15
相关论文
共 57 条
[1]  
Alicandri E., 2020, On the Evolution and Functional Diversity of Terpene Synthases in the Pinus Species: A Review
[2]   EnzyNet: enzyme classification using 3D convolutional neural networks on spatial representation [J].
Amidi, Afshine ;
Amidi, Shervine ;
Vlachakis, Dimitrios ;
Megalooikonomou, Vasileios ;
Paragios, Nikos ;
Zacharaki, Evangelia, I .
PEERJ, 2018, 6
[3]   Prediction and experimental validation of enzyme substrate specificity in protein structures [J].
Amin, Shivas R. ;
Erdin, Serkan ;
Ward, R. Matthew ;
Lua, Rhonald C. ;
Lichtarge, Olivier .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2013, 110 (45) :E4195-E4202
[4]  
Bello I., 2021, LambdaNetworks: Modeling Long-Range Interactions Without Attention
[5]  
Brandes N., 2022, ProteinBERT: a universal deep-learning model of protein sequence and function
[6]  
Cangea C, 2018, Arxiv, DOI [arXiv:1811.01287, DOI 10.48550/ARXIV.1811.01287]
[7]  
Christianson W.J.C.R., 2017, Structural and Chemical Biology of Terpenoid Cyclases, P11570
[8]   A Comprehensive Survey of Deep Learning Techniques in Protein Function Prediction [J].
Dhanuka, Richa ;
Singh, Jyoti Prakash ;
Tripathi, Anushree .
IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2023, 20 (03) :2291-2301
[9]  
Elabd H., 2020, Amino acid encoding for deep learning applications, P21
[10]   ProtTrans: Toward Understanding the Language of Life Through Self-Supervised Learning [J].
Elnaggar, Ahmed ;
Heinzinger, Michael ;
Dallago, Christian ;
Rehawi, Ghalia ;
Wang, Yu ;
Jones, Llion ;
Gibbs, Tom ;
Feher, Tamas ;
Angerer, Christoph ;
Steinegger, Martin ;
Bhowmik, Debsindhu ;
Rost, Burkhard .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) :7112-7127