Deep learning for NAD/NADP cofactor prediction and engineering using transformer attention analysis in enzymes

被引:0
|
作者
Kim, Jaehyung [1 ]
Woo, Jihoon [1 ]
Park, Joon Young [1 ]
Kim, Kyung-Jin [2 ]
Kim, Donghyuk [1 ]
机构
[1] Ulsan Natl Inst Sci & Technol UNIST, Sch Energy & Chem Engn, Ulsan 44919, South Korea
[2] Kyungpook Natl Univ, KNU Inst Microbiol, Sch Life Sci, BK21 FOUR KNU Creat Biores Grp, Daegu 41566, South Korea
基金
新加坡国家研究基金会;
关键词
NAD(P) specificity; Cofactor switching; Deep learning; Explainable AI; Protein engineering; Synthetic biology; COENZYME SPECIFICITY; REDUCTASE; BINDING; CLASSIFICATION; DEHYDROGENASE; PREFERENCE; PHOSPHATE; SUBSTRATE; SEQUENCE;
D O I
10.1016/j.ymben.2024.11.007
中图分类号
Q81 [生物工程学(生物技术)]; Q93 [微生物学];
学科分类号
071005 ; 0836 ; 090102 ; 100705 ;
摘要
Understanding and manipulating the cofactor preferences of NAD(P)-dependent oxidoreductases, the most widely distributed enzyme group in nature, is increasingly crucial in bioengineering. However, large-scale identification of the cofactor preferences and the design of mutants to switch cofactor specificity remain as complex tasks. Here, we introduce DISCODE (Deep learning-based Iterative pipeline to analyze Specificity of COfactors and to Design Enzyme), a novel transformer-based deep learning model to predict NAD(P) cofactor preferences. For model training, a total of 7,132 NAD(P)-dependent enzyme sequences were collected. Leveraging whole-length sequence information, DISCODE classifies the cofactor preferences of NAD(P)dependent oxidoreductase protein sequences without structural or taxonomic limitation. The model showed 97.4% and 97.3% of accuracy and F1 score, respectively. A notable feature of DISCODE is the interpretability of its transformer layers. Analysis of attention layers in the model enables identification of several residues that showed significantly higher attention weights. They were well aligned with structurally important residues that closely interact with NAD(P), facilitating the identification of key residues for determining cofactor specificities. These key residues showed high consistency with verified cofactor switching mutants. Integrated into an enzyme design pipeline, DISCODE coupled with attention analysis, enables a fully automated approach to redesign cofactor specificity.
引用
收藏
页码:86 / 94
页数:9
相关论文
共 50 条
  • [21] Deep learning analysis and age prediction from shoeprints
    Hassan, Muhammad
    Wang, Yan
    Wang, Di
    Li, Daixi
    Liang, Yanchun
    Zhou, You
    Xu, Dong
    FORENSIC SCIENCE INTERNATIONAL, 2021, 327
  • [22] Prediction of Heavy Rain Damage Using Deep Learning
    Lee, Kanghyeok
    Choi, Changhyun
    Shin, Do Hyoung
    Kim, Hung Soo
    WATER, 2020, 12 (07)
  • [23] Whole slide images based cancer survival prediction using attention guided deep multiple instance learning networks
    Yao, Jiawen
    Zhu, Xinliang
    Jonnagaddala, Jitendra
    Hawkins, Nicholas
    Huang, Junzhou
    MEDICAL IMAGE ANALYSIS, 2020, 65
  • [24] A Hybrid Attention-Based Transformer Model for Arabic News Classification Using Text Embedding and Deep Learning
    Hossain, Md. Mithun
    Hossain, Md. Shakil
    Safran, Mejdl
    Alfarhood, Sultan
    Alfarhood, Meshal
    F. Mridha, M.
    IEEE ACCESS, 2024, 12 : 198046 - 198066
  • [25] Construction of Transformer Fault Diagnosis and Prediction Model Based on Deep Learning
    Li X.
    Journal of Computing and Information Technology, 2022, 30 (04) : 223 - 238
  • [26] Transformer-Based Deep Learning Method for the Prediction of Ventilator Pressure
    Fan, Ruizhe
    2022 IEEE 2ND INTERNATIONAL CONFERENCE ON INFORMATION COMMUNICATION AND SOFTWARE ENGINEERING (ICICSE 2022), 2022, : 25 - 28
  • [27] Industrial Remaining Useful Life Prediction by Partial Observation Using Deep Learning With Supervised Attention
    Li, Xiang
    Jia, Xiaodong
    Wang, Yinglu
    Yang, Shaojie
    Zhao, Haodong
    Lee, Jay
    IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2020, 25 (05) : 2241 - 2251
  • [28] Argument annotation and analysis using deep learning with attention mechanism in Bahasa Indonesia
    Derwin Suhartono
    Aryo Pradipta Gema
    Suhendro Winton
    Theodorus David
    Mohamad Ivan Fanany
    Aniati Murni Arymurthy
    Journal of Big Data, 7
  • [29] Real-Time Driver's Focus of Attention Extraction and Prediction using Deep Learning
    Hong, Pei-heng
    Wang, Yuehua
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (06) : 1 - 10
  • [30] Deep Learning Framework Using Transformer Networks for Multi Building Energy Consumption Prediction in Smart Cities
    Moveh, Samuel
    Merchan-Cruz, Emmanuel Alejandro
    Abuhussain, Maher
    Dodo, Yakubu Aminu
    Alhumaid, Saleh
    Alhamami, Ali Hussain
    ENERGIES, 2025, 18 (06)