Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks

被引:14
|
作者
Luo, Ling [1 ,2 ]
Ning, Jinzhong [1 ]
Zhao, Yingwen [1 ]
Wang, Zhijun [1 ]
Ding, Zeyuan [1 ]
Chen, Peng [1 ]
Fu, Weiru [1 ]
Han, Qinyu [1 ]
Xu, Guangtao [1 ]
Qiu, Yunzhi [1 ]
Pan, Dinghao [1 ]
Li, Jiru [1 ]
Li, Hao [1 ]
Feng, Wenduo [1 ]
Tu, Senbo [1 ]
Liu, Yuqi [1 ]
Yang, Zhihao [1 ]
Wang, Jian [1 ]
Sun, Yuanyuan [1 ]
Lin, Hongfei [1 ]
机构
[1] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian 116024, Peoples R China
[2] Dalian Univ Technol, Sch Comp Sci & Technol, 2 Linggong Rd, Ganjingzi Dist, Dalian 116024, Peoples R China
基金
中国国家自然科学基金;
关键词
natural language processing; large language model; supervised fine-tuning; biomedical multitasking;
D O I
10.1093/jamia/ocae037
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objective Most existing fine-tuned biomedical large language models (LLMs) focus on enhancing performance in monolingual biomedical question answering and conversation tasks. To investigate the effectiveness of the fine-tuned LLMs on diverse biomedical natural language processing (NLP) tasks in different languages, we present Taiyi, a bilingual fine-tuned LLM for diverse biomedical NLP tasks.Materials and Methods We first curated a comprehensive collection of 140 existing biomedical text mining datasets (102 English and 38 Chinese datasets) across over 10 task types. Subsequently, these corpora were converted to the instruction data used to fine-tune the general LLM. During the supervised fine-tuning phase, a 2-stage strategy is proposed to optimize the model performance across various tasks.Results Experimental results on 13 test sets, which include named entity recognition, relation extraction, text classification, and question answering tasks, demonstrate that Taiyi achieves superior performance compared to general LLMs. The case study involving additional biomedical NLP tasks further shows Taiyi's considerable potential for bilingual biomedical multitasking.Conclusion Leveraging rich high-quality biomedical corpora and developing effective fine-tuning strategies can significantly improve the performance of LLMs within the biomedical domain. Taiyi shows the bilingual multitasking capability through supervised fine-tuning. However, those tasks such as information extraction that are not generation tasks in nature remain challenging for LLM-based generative approaches, and they still underperform the conventional discriminative approaches using smaller language models.
引用
收藏
页码:1865 / 1874
页数:10
相关论文
共 50 条
  • [41] Mining Insights from Large-Scale Corpora Using Fine-Tuned Language Models
    Palakodety, Shriphani
    KhudaBukhsh, Ashiqur R.
    Carbonell, Jaime G.
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1890 - 1897
  • [42] On the Generalization Abilities of Fine-Tuned Commonsense Language Representation Models
    Shen, Ke
    Kejriwal, Mayank
    ARTIFICIAL INTELLIGENCE XXXVIII, 2021, 13101 : 3 - 16
  • [43] Deciphering language disturbances in schizophrenia: A study using fine-tuned language models
    Li, Renyu
    Cao, Minne
    Fu, Dawei
    Wei, Wei
    Wang, Dequan
    Yuan, Zhaoxia
    Hu, Ruofei
    Deng, Wei
    SCHIZOPHRENIA RESEARCH, 2024, 271 : 120 - 128
  • [44] ChatDoctor: A Medical Chat Model Fine-Tuned on a Large Language Model Meta-AI (LLaMA) Using Medical Domain Knowledge
    Li, Yunxiang
    Li, Zihan
    Zhang, Kai
    Dan, Ruilong
    Jiang, Steve
    Zhang, You
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (06)
  • [45] PharmBERT: a Fine-tuned Model for Pharmaceutical Error Prediction
    Hu, Gang
    Yu, Bo
    Doctor, Dustin
    2023 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI, 2023, : 343 - 344
  • [46] Softly fine-tuned Standard Model and the scale of inflation
    Korutlu, Beste
    MODERN PHYSICS LETTERS A, 2015, 30 (34)
  • [47] How fine-tuned is a large Muon EDM from Flavor?
    Ruppell, Timo
    SUSY09: THE 17TH INTERNATIONAL CONFERENCE ON SUPERSYMMETRY AND THE UNIFICATION OF FUNDAMENTAL INTERACTIONS, 2009, 1200 : 900 - 903
  • [48] An open-source fine-tuned large language model for radiological impression generation: a multi-reader performance study
    Serapio, Adrian
    Chaudhari, Gunvant
    Savage, Cody
    Lee, Yoo Jin
    Vella, Maya
    Sridhar, Shravan
    Schroeder, Jamie Lee
    Liu, Jonathan
    Yala, Adam
    Sohn, Jae Ho
    BMC MEDICAL IMAGING, 2024, 24 (01):
  • [49] Assessment of fine-tuned large language models for real-world chemistry and material science applications
    Van Herck, Joren
    Gil, Maria Victoria
    Jablonka, Kevin Maik
    Abrudan, Alex
    Anker, Andy S.
    Asgari, Mehrdad
    Blaiszik, Ben
    Buffo, Antonio
    Choudhury, Leander
    Corminboeuf, Clemence
    Daglar, Hilal
    Elahi, Amir Mohammad
    Foster, Ian T.
    Garcia, Susana
    Garvin, Matthew
    Godin, Guillaume
    Good, Lydia L.
    Gu, Jianan
    Xiao Hu, Noemie
    Jin, Xin
    Junkers, Tanja
    Keskin, Seda
    Knowles, Tuomas P. J.
    Laplaza, Ruben
    Lessona, Michele
    Majumdar, Sauradeep
    Mashhadimoslem, Hossein
    Mcintosh, Ruaraidh D.
    Moosavi, Seyed Mohamad
    Mourino, Beatriz
    Nerli, Francesca
    Pevida, Covadonga
    Poudineh, Neda
    Rajabi-Kochi, Mahyar
    Saar, Kadi L.
    Hooriabad Saboor, Fahimeh
    Sagharichiha, Morteza
    Schmidt, K. J.
    Shi, Jiale
    Simone, Elena
    Svatunek, Dennis
    Taddei, Marco
    Tetko, Igor
    Tolnai, Domonkos
    Vahdatifar, Sahar
    Whitmer, Jonathan
    Wieland, D. C. Florian
    Willumeit-Roemer, Regine
    Zuttel, Andreas
    Smit, Berend
    CHEMICAL SCIENCE, 2025, 16 (02) : 670 - 684
  • [50] Enhancing Zero-Shot Crypto Sentiment With Fine-Tuned Language Model and Prompt Engineering
    Wahidur, Rahman S. M.
    Tashdeed, Ishmam
    Kaur, Manjit
    Lee, Heung-No
    IEEE ACCESS, 2024, 12 : 10146 - 10159