Leveraging pre-trained language models for mining microbiome-disease relationships

被引:12
作者
Karkera, Nikitha [4 ]
Acharya, Sathwik [1 ,3 ]
Palaniappan, Sucheendra K. [1 ,2 ,4 ]
机构
[1] Syst Biol Inst, Tokyo, Japan
[2] Iom Bioworks Pvt Ltd, Bengaluru, India
[3] PES Univ, Bengaluru, India
[4] SBX Corp, Tokyo, Japan
关键词
Microbe-disease relationship extraction; Language models; Fine-tuning; Deep-learning; Transfer learning; Biomedical informatics; Natural language processing; DATABASE;
D O I
10.1186/s12859-023-05411-z
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
Background: The growing recognition of the microbiome's impact on human health and well-being has prompted extensive research into discovering the links between microbiome dysbiosis and disease (healthy) states. However, this valuable information is scattered in unstructured form within biomedical literature. The struc-tured extraction and qualification of microbe-disease interactions are important. In parallel, recent advancements in deep-learning-based natural language processing algorithms have revolutionized language-related tasks such as ours. This study aims to leverage state-of-the-art deep-learning language models to extract microbe-disease relationships from biomedical literature.Results: In this study, we first evaluate multiple pre-trained large language models within a zero-shot or few-shot learning context. In this setting, the models performed poorly out of the box, emphasizing the need for domain-specific fine-tuning of these language models. Subsequently, we fine-tune multiple language models (specifi-cally, GPT-3, BioGPT, BioMedLM, BERT, BioMegatron, PubMedBERT, BioClinicalBERT, and BioLinkBERT) using labeled training data and evaluate their performance. Our experimental results demonstrate the state-of-the-art performance of these fine-tuned models ( specifically GPT-3, BioMedLM, and BioLinkBERT), achieving an average F1 score, precision, and recall of over > 0.8 compared to the previous best of 0.74.Conclusion: Overall, this study establishes that pre-trained language models excel as transfer learners when fine-tuned with domain and problem-specific data, enabling them to achieve state-of-the-art results even with limited training data for extracting microbiome-disease interactions from scientific publications.
引用
收藏
页数:19
相关论文
共 49 条
  • [1] Achiam OJ, 2023, Arxiv, DOI [arXiv:2303.08774, DOI 10.48550/ARXIV.2303.08774]
  • [2] Ahmed SAJA, 2022, FRONT PHYSIOL, V13
  • [3] Alsentzer E, 2019, Arxiv, DOI [arXiv:1904.03323, DOI 10.48550/ARXIV.1904.03323]
  • [4] Challenges in the construction of knowledge bases for human microbiome-disease associations
    Badal, Varsha Dave
    Wright, Dustin
    Katsis, Yannis
    Kim, Ho-Cheol
    Swafford, Austin D.
    Knight, Rob
    Hsu, Chun-Nan
    [J]. MICROBIOME, 2019, 7 (01)
  • [5] Novel human microbe-disease association prediction using network consistency projection
    Bao, Wenzheng
    Jiang, Zhichao
    Huang, De-Shuang
    [J]. BMC BIOINFORMATICS, 2017, 18
  • [6] Brown T., 2020, ADV NEURAL INFORM PR, P1877
  • [7] gutMDisorder: a comprehensive database for dysbiosis of the gut microbiota in disorders and interventions
    Cheng, Liang
    Qi, Changlu
    Zhuang, He
    Fu, Tongze
    Zhang, Xue
    [J]. NUCLEIC ACIDS RESEARCH, 2020, 48 (D1) : D554 - D560
  • [8] Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
  • [9] Gu Y, 2020, arXiv
  • [10] A novel machine learning framework for automated biomedical relation extraction from large-scale literature repositories
    Hong, Lixiang
    Lin, Jinjian
    Li, Shuya
    Wan, Fangping
    Yang, Hui
    Jiang, Tao
    Zhao, Dan
    Zeng, Jianyang
    [J]. NATURE MACHINE INTELLIGENCE, 2020, 2 (06) : 347 - +