PreAlgPro: Prediction of allergenic proteins with pre-trained protein language model and efficient neutral network

被引:1
|
作者
Zhang, Lingrong [1 ]
Liu, Taigang [1 ]
机构
[1] Shanghai Ocean Univ, Coll Informat Technol, Shanghai 201306, Peoples R China
关键词
Pre-trained protein language model; Allergenic proteins; Deep learning; Model interpretability; DATABASE;
D O I
10.1016/j.ijbiomac.2024.135762
中图分类号
Q5 [生物化学]; Q7 [分子生物学];
学科分类号
071010 ; 081704 ;
摘要
Allergy is a prevalent phenomenon, involving allergens such as nuts and milk. Avoiding exposure to allergens is the most effective preventive measure against allergic reactions. However, current homology-based methods for identifying allergenic proteins encounter challenges when dealing with non-homologous data. Traditional machine learning approaches rely on manually extracted features, which lack important protein functional characteristics, including evolutionary information. Consequently, there is still considerable room for improvement in existing methods. In this study, we present PreAlgPro, a method for identifying allergenic proteins based on pre-trained protein language models and deep learning techniques. Specifically, we employed the ProtT5 model to extract protein embedding features, replacing the manual feature extraction step. Furthermore, we devised an Attention-CNN neural network architecture to identify potential features that contribute to the classification of allergenic proteins. The performance of our model was evaluated on four independent test sets, and the experimental results demonstrate that PreAlgPro surpasses existing state-of-the-art methods. Additionally, we collected allergenic protein samples to validate the robustness of the model and conducted an analysis of model interpretability.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Biomedical generative pre-trained based transformer language model for age-related disease target discovery
    Zagirova, Diana
    Pushkov, Stefan
    Leung, Geoffrey Ho Duen
    Liu, Bonnie Hei Man
    Urban, Anatoly
    Sidorenko, Denis
    Kalashnikov, Aleksandr
    Kozlova, Ekaterina
    Naumov, Vladimir
    Pun, Frank W.
    Ozerov, Ivan V.
    Aliper, Alex
    Zhavoronkov, Alex
    AGING-US, 2023, 15 (18): : 9293 - 9309
  • [32] VesiMCNN: Using pre-trained protein language models and multiple window scanning convolutional neural networks to identify vesicular transport proteins
    Le, Van The
    Tseng, Yi-Hsuan
    Liu, Yu-Chen
    Malik, Muhammad Shahid
    Ou, Yu-Yen
    INTERNATIONAL JOURNAL OF BIOLOGICAL MACROMOLECULES, 2024, 280
  • [33] Improving automatic cyberbullying detection in social network environments by fine-tuning a pre-trained sentence transformer language model
    Gutierrez-Batista, Karel
    Gomez-Sanchez, Jesica
    Fernandez-Basso, Carlos
    SOCIAL NETWORK ANALYSIS AND MINING, 2024, 14 (01)
  • [34] Pre-Trained Language Model-Based Deep Learning for Sentiment Classification of Vietnamese Feedback
    Loc, Cu Vinh
    Viet, Truong Xuan
    Viet, Tran Hoang
    Thao, Le Hoang
    Viet, Nguyen Hoang
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2023, 22 (03)
  • [35] Basic investigation of sign language motion classification by feature extraction using pre-trained network models
    Kawaguchi, Kaito
    Nishimura, Hiromitsu
    Wang, Zhizhong
    Tanaka, Hiroshi
    Ohta, Eiji
    2019 IEEE PACIFIC RIM CONFERENCE ON COMMUNICATIONS, COMPUTERS AND SIGNAL PROCESSING (PACRIM), 2019,
  • [36] T4SEfinder: a bioinformatics tool for genome-scale prediction of bacterial type IV secreted effectors using pre-trained protein language model
    Zhang, Yumeng
    Zhang, Yangming
    Xiong, Yi
    Wang, Hui
    Deng, Zixin
    Song, Jiangning
    Ou, Hong-Yu
    BRIEFINGS IN BIOINFORMATICS, 2022, 23 (01)
  • [37] Transfer-DDG: Prediction of protein-protein binding affinity changes with mutations based on large pre-trained model transfer learning
    Wang, Yuxiang
    Shi, Xiumin
    Zhou, Han
    2023 IEEE 2ND INDUSTRIAL ELECTRONICS SOCIETY ANNUAL ON-LINE CONFERENCE, ONCON, 2023,
  • [38] Rapid seismic damage state prediction of the subway station structure using the pre-trained network and convolutional neural network
    Fan, Yifan
    Chen, Zhiyi
    Luo, Xiaowei
    SOIL DYNAMICS AND EARTHQUAKE ENGINEERING, 2024, 185
  • [39] Transfer Learning Effects on Image Steganalysis with Pre-Trained Deep Residual Neural Network Model
    Ozcan, Selim
    Mustacoglu, Ahmet Fatih
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 2280 - 2287
  • [40] PEPC: A Deep Parallel Convolutional Neural Network Model with Pre-trained Embeddings for DGA Detection
    Huang, Weiqing
    Zong, Yangyang
    Shi, Zhixin
    Wang, Leiqi
    Liu, Pengcheng
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,