Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks

被引:14
|
作者
Luo, Ling [1 ,2 ]
Ning, Jinzhong [1 ]
Zhao, Yingwen [1 ]
Wang, Zhijun [1 ]
Ding, Zeyuan [1 ]
Chen, Peng [1 ]
Fu, Weiru [1 ]
Han, Qinyu [1 ]
Xu, Guangtao [1 ]
Qiu, Yunzhi [1 ]
Pan, Dinghao [1 ]
Li, Jiru [1 ]
Li, Hao [1 ]
Feng, Wenduo [1 ]
Tu, Senbo [1 ]
Liu, Yuqi [1 ]
Yang, Zhihao [1 ]
Wang, Jian [1 ]
Sun, Yuanyuan [1 ]
Lin, Hongfei [1 ]
机构
[1] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian 116024, Peoples R China
[2] Dalian Univ Technol, Sch Comp Sci & Technol, 2 Linggong Rd, Ganjingzi Dist, Dalian 116024, Peoples R China
基金
中国国家自然科学基金;
关键词
natural language processing; large language model; supervised fine-tuning; biomedical multitasking;
D O I
10.1093/jamia/ocae037
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objective Most existing fine-tuned biomedical large language models (LLMs) focus on enhancing performance in monolingual biomedical question answering and conversation tasks. To investigate the effectiveness of the fine-tuned LLMs on diverse biomedical natural language processing (NLP) tasks in different languages, we present Taiyi, a bilingual fine-tuned LLM for diverse biomedical NLP tasks.Materials and Methods We first curated a comprehensive collection of 140 existing biomedical text mining datasets (102 English and 38 Chinese datasets) across over 10 task types. Subsequently, these corpora were converted to the instruction data used to fine-tune the general LLM. During the supervised fine-tuning phase, a 2-stage strategy is proposed to optimize the model performance across various tasks.Results Experimental results on 13 test sets, which include named entity recognition, relation extraction, text classification, and question answering tasks, demonstrate that Taiyi achieves superior performance compared to general LLMs. The case study involving additional biomedical NLP tasks further shows Taiyi's considerable potential for bilingual biomedical multitasking.Conclusion Leveraging rich high-quality biomedical corpora and developing effective fine-tuning strategies can significantly improve the performance of LLMs within the biomedical domain. Taiyi shows the bilingual multitasking capability through supervised fine-tuning. However, those tasks such as information extraction that are not generation tasks in nature remain challenging for LLM-based generative approaches, and they still underperform the conventional discriminative approaches using smaller language models.
引用
收藏
页码:1865 / 1874
页数:10
相关论文
共 50 条
  • [21] Automated classification of brain MRI reports using fine-tuned large language models
    Kanzawa, Jun
    Yasaka, Koichiro
    Fujita, Nana
    Fujiwara, Shin
    Abe, Osamu
    NEURORADIOLOGY, 2024, 66 (12) : 2177 - 2183
  • [22] Efficacy of Fine-Tuned Large Language Model in CT Protocol Assignment as Clinical Decision-Supporting System
    Kanemaru, Noriko
    Yasaka, Koichiro
    Okimoto, Naomasa
    Sato, Mai
    Nomura, Takuto
    Morita, Yuichi
    Katayama, Akira
    Kiryu, Shigeru
    Abe, Osamu
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025,
  • [23] The Impact of AUTOGEN and Similar Fine-Tuned Large Language Models on the Integrity of Scholarly Writing
    Resnik, David B.
    Hosseini, Mohammad
    AMERICAN JOURNAL OF BIOETHICS, 2023, 23 (10): : 50 - 52
  • [24] Comparative Analysis of Generic and Fine-Tuned Large Language Models for Conversational Agent Systems
    Villa, Laura
    Carneros-Prado, David
    Dobrescu, Cosmin C.
    Sanchez-Miguel, Adrian
    Cubero, Guillermo
    Hervas, Ramon
    ROBOTICS, 2024, 13 (05)
  • [25] Understanding language-elicited EEG data by predicting it from a fine-tuned language model
    Schwartz, Dan
    Mitchell, Tom
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 43 - 57
  • [26] Fine-Tuned Transformer Model for Sentiment Analysis
    Liu, Sishun
    Shuai, Pengju
    Zhang, Xiaowu
    Chen, Shuang
    Li, Li
    Liu, Ming
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT (KSEM 2020), PT II, 2020, 12275 : 336 - 343
  • [27] A minimally fine-tuned supersymmetric standard model
    Chacko, Z
    Nomura, Y
    Tucker-Smith, D
    NUCLEAR PHYSICS B, 2005, 725 (1-2) : 207 - 250
  • [28] A generalist vision-language foundation model for diverse biomedical tasks
    Zhang, Kai
    Zhou, Rong
    Adhikarla, Eashan
    Yan, Zhiling
    Liu, Yixin
    Yu, Jun
    Liu, Zhengliang
    Chen, Xun
    Davison, Brian D.
    Ren, Hui
    Huang, Jing
    Chen, Chen
    Zhou, Yuyin
    Fu, Sunyang
    Liu, Wei
    Liu, Tianming
    Li, Xiang
    Chen, Yong
    He, Lifang
    Zou, James
    Li, Quanzheng
    Liu, Hongfang
    Sun, Lichao
    NATURE MEDICINE, 2024, 30 (11) : 3129 - 3141
  • [29] CD-LLMCARS: Cross Domain Fine-Tuned Large Language Model for Context-Aware Recommender Systems
    Cheema, Adeel Ashraf
    Sarfraz, Muhammad Shahzad
    Habib, Usman
    Zaman, Qamar Uz
    Boonchieng, Ekkarat
    IEEE OPEN JOURNAL OF THE COMPUTER SOCIETY, 2025, 6 : 49 - 59
  • [30] NM-GPT: Advancing Nuclear Medicine Report Processing Through a Specialized Fine-tuned Large Language Model
    Lyu, Zhiliang
    Zeng, Fang
    Guo, Ning
    Li, Xiang
    Li, Quanzheng
    JOURNAL OF NUCLEAR MEDICINE, 2024, 65