AMP-BERT: Prediction of antimicrobial peptide function based on a BERT model

被引:51
作者
Lee, Hansol [1 ]
Lee, Songyeon [1 ]
Lee, Ingoo [1 ]
Nam, Hojung [1 ,2 ]
机构
[1] Gwangju Inst Sci & Technol GIST, Sch Elect Engn & Comp Sci, Gwangju, South Korea
[2] Gwangju Inst Sci & Technol, AI Grad Sch, 123 Cheomdangwagi Ro, Gwangju 61005, South Korea
基金
新加坡国家研究基金会;
关键词
antimicrobial peptides; antimicrobial resistance; BERT; deep learning; drug discovery; machine learning; sequence classification; transformer; CD-HIT; PROTEIN; THIONINS;
D O I
10.1002/pro.4529
中图分类号
Q5 [生物化学]; Q7 [分子生物学];
学科分类号
071010 ; 081704 ;
摘要
Antimicrobial resistance is a growing health concern. Antimicrobial peptides (AMPs) disrupt harmful microorganisms by nonspecific mechanisms, making it difficult for microbes to develop resistance. Accordingly, they are promising alternatives to traditional antimicrobial drugs. In this study, we developed an improved AMP classification model, called AMP-BERT. We propose a deep learning model with a fine-tuned didirectional encoder representations from transformers (BERT) architecture designed to extract structural/functional information from input peptides and identify each input as AMP or non-AMP. We compared the performance of our proposed model and other machine/deep learning-based methods. Our model, AMP-BERT, yielded the best prediction results among all models evaluated with our curated external dataset. In addition, we utilized the attention mechanism in BERT to implement an interpretable feature analysis and determine the specific residues in known AMPs that contribute to peptide structure and antimicrobial function. The results show that AMP-BERT can capture the structural properties of peptides for model learning, enabling the prediction of AMPs or non-AMPs from input sequences. AMP-BERT is expected to contribute to the identification of candidate AMPs for functional validation and drug development. The code and dataset for the fine-tuning of AMP-BERT is publicly available at .
引用
收藏
页数:13
相关论文
共 50 条
[21]   BERT-TFBS: a novel BERT-based model for predicting transcription factor binding sites by transfer learning [J].
Wang, Kai ;
Zeng, Xuan ;
Zhou, Jingwen ;
Liu, Fei ;
Luan, Xiaoli ;
Wang, Xinglong .
BRIEFINGS IN BIOINFORMATICS, 2024, 25 (03)
[22]   A BERT-based model for the prediction of lncRNA subcellular localization in Homo sapiens [J].
Zhang, Zhao-Yue ;
Zhang, Zheng ;
Ye, Xiucai ;
Sakurai, Tetsuya ;
Lin, Hao .
INTERNATIONAL JOURNAL OF BIOLOGICAL MACROMOLECULES, 2024, 265
[23]   Automatic Prediction and Linguistic Interpretation of Chinese Directional Complements Based on BERT Model [J].
Jeong, Young Hoon ;
Li, Ming Yue ;
Kang, Su Min ;
Eum, Yun Kyung ;
Kang, Byeong Kwu .
CHINESE LEXICAL SEMANTICS, CLSW 2021, PT I, 2022, 13249 :405-416
[24]   Adaptive Thresholding for Sentiment Analysis Across Online Reviews Based on BERT Model BERT-based Adaptive Thresholding for Sentiment Analysis [J].
Lu, Zijie .
PROCEEDINGS OF INTERNATIONAL CONFERENCE ON MODELING, NATURAL LANGUAGE PROCESSING AND MACHINE LEARNING, CMNM 2024, 2024, :70-75
[25]   BERT based severity prediction of bug reports for the maintenance of mobile applications [J].
Ali, Asif ;
Xia, Yuanqing ;
Umer, Qasim ;
Osman, Mohamed .
JOURNAL OF SYSTEMS AND SOFTWARE, 2024, 208
[26]   A BERT-based Idiom Detection Model [J].
Gamage, Gihan ;
De Silva, Daswin ;
Adikari, Achini ;
Alahakoon, Damminda .
2022 15TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION (HSI), 2022,
[27]   Chinese Triple Extraction Based on BERT Model [J].
Deng, Weidong ;
Liu, Yun .
PROCEEDINGS OF THE 2021 15TH INTERNATIONAL CONFERENCE ON UBIQUITOUS INFORMATION MANAGEMENT AND COMMUNICATION (IMCOM 2021), 2021,
[28]   BERT-based keyword extraction model for the Turkish language [J].
Babayigit, Bilal ;
Sattuf, Hamza .
Neural Computing and Applications, 2025, 37 (16) :9807-9819
[29]   Chinese Text Sentiment Analysis Model Based on BERT and BiTCN [J].
Chen, Jinlan ;
Zhang, Jian ;
Zhang, Jiajing ;
Chen, Shufeng .
2024 17TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTER THEORY AND ENGINEERING, ICACTE, 2024, :127-133
[30]   Transformer based Contextual Model for Sentiment Analysis of Customer Reviews: A Fine-tuned BERT A Sequence Learning BERT Model for Sentiment Analysis [J].
Durairaj, Ashok Kumar ;
Chinnalagu, Anandan .
INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2021, 12 (11) :474-480