AMP-BERT: Prediction of antimicrobial peptide function based on a BERT model

被引:39
|
作者
Lee, Hansol [1 ]
Lee, Songyeon [1 ]
Lee, Ingoo [1 ]
Nam, Hojung [1 ,2 ]
机构
[1] Gwangju Inst Sci & Technol GIST, Sch Elect Engn & Comp Sci, Gwangju, South Korea
[2] Gwangju Inst Sci & Technol, AI Grad Sch, 123 Cheomdangwagi Ro, Gwangju 61005, South Korea
基金
新加坡国家研究基金会;
关键词
antimicrobial peptides; antimicrobial resistance; BERT; deep learning; drug discovery; machine learning; sequence classification; transformer; CD-HIT; PROTEIN; THIONINS;
D O I
10.1002/pro.4529
中图分类号
Q5 [生物化学]; Q7 [分子生物学];
学科分类号
071010 ; 081704 ;
摘要
Antimicrobial resistance is a growing health concern. Antimicrobial peptides (AMPs) disrupt harmful microorganisms by nonspecific mechanisms, making it difficult for microbes to develop resistance. Accordingly, they are promising alternatives to traditional antimicrobial drugs. In this study, we developed an improved AMP classification model, called AMP-BERT. We propose a deep learning model with a fine-tuned didirectional encoder representations from transformers (BERT) architecture designed to extract structural/functional information from input peptides and identify each input as AMP or non-AMP. We compared the performance of our proposed model and other machine/deep learning-based methods. Our model, AMP-BERT, yielded the best prediction results among all models evaluated with our curated external dataset. In addition, we utilized the attention mechanism in BERT to implement an interpretable feature analysis and determine the specific residues in known AMPs that contribute to peptide structure and antimicrobial function. The results show that AMP-BERT can capture the structural properties of peptides for model learning, enabling the prediction of AMPs or non-AMPs from input sequences. AMP-BERT is expected to contribute to the identification of candidate AMPs for functional validation and drug development. The code and dataset for the fine-tuning of AMP-BERT is publicly available at .
引用
收藏
页数:13
相关论文
共 50 条
  • [11] AFS-BERT: Information entropy-based adaptive fusion sampling and Bert embedding model for link prediction
    Zhang, Lei
    Pan, Jiaxing
    Ma, Xiaoxuan
    Yang, Chengwei
    INTERNATIONAL JOURNAL OF MODERN PHYSICS B, 2023,
  • [12] EHR-BERT: A BERT-based model for effective anomaly detection in electronic health records
    Niu, Haoran
    Omitaomu, Olufemi A.
    Langston, Michael A.
    Olama, Mohammad
    Ozmen, Ozgur
    Klasky, Hilda B.
    Laurio, Angela
    Ward, Merry
    Nebeker, Jonathan
    JOURNAL OF BIOMEDICAL INFORMATICS, 2024, 150
  • [13] A BERT-based review helpfulness prediction model utilizing consistency of ratings and texts
    Li, Xinzhe
    Li, Qinglong
    Ryu, Dongyeop
    Kim, Jaekyeong
    APPLIED INTELLIGENCE, 2025, 55 (06)
  • [14] NABP-BERT: NANOBODY®-antigen binding prediction based on bidirectional encoder representations from transformers (BERT) architecture
    Ahmed, Fatma S.
    Aly, Saleh
    Liu, Xiangrong
    BRIEFINGS IN BIOINFORMATICS, 2024, 26 (01)
  • [15] Deep Learning Algorithm for Judicial Judgment Prediction Based on BERT
    Wang, Yongjun
    Gao, Jing
    Chen, Junjie
    PROCEEDINGS OF THE 2020 5TH INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION AND SECURITY (ICCCS-2020), 2020,
  • [16] Comprehensive Assessment of BERT-Based Methods for Predicting Antimicrobial Peptides
    Gao, Wanling
    Zhao, Jun
    Gui, Jianfeng
    Wang, Zehan
    Chen, Jie
    Yue, Zhenyu
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2024, 64 (19) : 7772 - 7785
  • [17] iAMP-Attenpred: a novel antimicrobial peptide predictor based on BERT feature extraction method and CNN-BiLSTM-Attention combination model
    Xing, Wenxuan
    Zhang, Jie
    Li, Chen
    Huo, Yujia
    Dong, Gaifang
    BRIEFINGS IN BIOINFORMATICS, 2024, 25 (01)
  • [18] BERT-5mC: an interpretable model for predicting 5-methylcytosine sites of DNA based on BERT
    Wang, Shuyu
    Liu, Yinbo
    Liu, Yufeng
    Zhang, Yong
    Zhu, Xiaolei
    PEERJ, 2023, 11
  • [19] Automatic Prediction and Linguistic Interpretation of Chinese Directional Complements Based on BERT Model
    Jeong, Young Hoon
    Li, Ming Yue
    Kang, Su Min
    Eum, Yun Kyung
    Kang, Byeong Kwu
    CHINESE LEXICAL SEMANTICS, CLSW 2021, PT I, 2022, 13249 : 405 - 416
  • [20] A BERT-based model for the prediction of lncRNA subcellular localization in Homo sapiens
    Zhang, Zhao-Yue
    Zhang, Zheng
    Ye, Xiucai
    Sakurai, Tetsuya
    Lin, Hao
    INTERNATIONAL JOURNAL OF BIOLOGICAL MACROMOLECULES, 2024, 265