Taiyi: a bilingual fine-tuned large language model for diverse biomedical tasks

被引:14
|
作者
Luo, Ling [1 ,2 ]
Ning, Jinzhong [1 ]
Zhao, Yingwen [1 ]
Wang, Zhijun [1 ]
Ding, Zeyuan [1 ]
Chen, Peng [1 ]
Fu, Weiru [1 ]
Han, Qinyu [1 ]
Xu, Guangtao [1 ]
Qiu, Yunzhi [1 ]
Pan, Dinghao [1 ]
Li, Jiru [1 ]
Li, Hao [1 ]
Feng, Wenduo [1 ]
Tu, Senbo [1 ]
Liu, Yuqi [1 ]
Yang, Zhihao [1 ]
Wang, Jian [1 ]
Sun, Yuanyuan [1 ]
Lin, Hongfei [1 ]
机构
[1] Dalian Univ Technol, Sch Comp Sci & Technol, Dalian 116024, Peoples R China
[2] Dalian Univ Technol, Sch Comp Sci & Technol, 2 Linggong Rd, Ganjingzi Dist, Dalian 116024, Peoples R China
基金
中国国家自然科学基金;
关键词
natural language processing; large language model; supervised fine-tuning; biomedical multitasking;
D O I
10.1093/jamia/ocae037
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objective Most existing fine-tuned biomedical large language models (LLMs) focus on enhancing performance in monolingual biomedical question answering and conversation tasks. To investigate the effectiveness of the fine-tuned LLMs on diverse biomedical natural language processing (NLP) tasks in different languages, we present Taiyi, a bilingual fine-tuned LLM for diverse biomedical NLP tasks.Materials and Methods We first curated a comprehensive collection of 140 existing biomedical text mining datasets (102 English and 38 Chinese datasets) across over 10 task types. Subsequently, these corpora were converted to the instruction data used to fine-tune the general LLM. During the supervised fine-tuning phase, a 2-stage strategy is proposed to optimize the model performance across various tasks.Results Experimental results on 13 test sets, which include named entity recognition, relation extraction, text classification, and question answering tasks, demonstrate that Taiyi achieves superior performance compared to general LLMs. The case study involving additional biomedical NLP tasks further shows Taiyi's considerable potential for bilingual biomedical multitasking.Conclusion Leveraging rich high-quality biomedical corpora and developing effective fine-tuning strategies can significantly improve the performance of LLMs within the biomedical domain. Taiyi shows the bilingual multitasking capability through supervised fine-tuning. However, those tasks such as information extraction that are not generation tasks in nature remain challenging for LLM-based generative approaches, and they still underperform the conventional discriminative approaches using smaller language models.
引用
收藏
页码:1865 / 1874
页数:10
相关论文
共 50 条
  • [31] Multi-class hate speech detection in the Norwegian language using FAST-RNN and multilingual fine-tuned transformers
    Ehtesham Hashmi
    Sule Yildirim Yayilgan
    Complex & Intelligent Systems, 2024, 10 : 4535 - 4556
  • [32] Contextual semantic embeddings based on fine-tuned AraBERT model for Arabic text multi-class categorization
    El-Alami, Fatima-zahra
    El Alaoui, Said Ouatik
    Nahnahi, Noureddine En
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (10) : 8422 - 8428
  • [33] Structuring medication signeturs as a language regression task: comparison of zero- and few-shot GPT with fine-tuned models
    Garcia-Agundez, Augusto
    Kay, Julia L.
    Li, Jing
    Gianfrancesco, Milena
    Rai, Baljeet
    Hu, Angela
    Schmajuk, Gabriela
    Yazdany, Jinoos
    JAMIA OPEN, 2024, 7 (02)
  • [34] FinBERT-FOMC: Fine-Tuned FinBERT Model with Sentiment Focus Method for Enhancing Sentiment Analysis of FOMC Minutes
    Chen, Ziwei
    Goessi, Sandro
    Kim, Wonseong
    Bermeitinger, Bernhard
    Handschuh, Siegfried
    PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2023, 2023, : 357 - 364
  • [35] MediBioDeBERTa: Biomedical Language Model With Continuous Learning and Intermediate Fine-Tuning
    Kim, Eunhui
    Jeong, Yuna
    Choi, Myung-Seok
    IEEE ACCESS, 2023, 11 : 141036 - 141044
  • [36] A Fine-Tuned Bidirectional Encoder Representations From Transformers Model for Food Named-Entity Recognition: Algorithm Development and Validation
    Stojanov, Riste
    Popovski, Gorjan
    Cenikj, Gjorgjina
    Seljak, Barbara Korousic
    Eftimov, Tome
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2021, 23 (08)
  • [37] Fine tuning the large language pegasus model for dialogue summarization
    Vinay Sarthak
    Preeti Rishiwal
    Mano Yadav
    Sushil Yadav
    Ashutosh Gangwar
    undefined Shankdhar
    International Journal of Information Technology, 2025, 17 (2) : 1165 - 1177
  • [38] Exploring Public Attitude Towards Children by Leveraging Emoji to Track Out Sentiment Using Distil-BERT a Fine-Tuned Model
    Saha, Uchchhwas
    Mahmud, Md. Shihab
    Keya, Mumenunnessa
    Lucky, Effat Ara Easmin
    Khushbu, Sharun Akter
    Noori, Sheak Rashed Haider
    Syed, Muntaser Mansur
    THIRD INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND CAPSULE NETWORKS (ICIPCN 2022), 2022, 514 : 332 - 346
  • [39] Similarity-Based Prompt Construction for Large Language Model in Medical Tasks
    Liu, Gaofei
    Pan, Meiqi
    Ma, Zhiyuan
    Gu, Miaomiao
    Yang, Ling
    Qin, Jiwei
    HEALTH INFORMATION PROCESSING: EVALUATION TRACK PAPERS, CHIP 2023, 2024, 2080 : 73 - 83
  • [40] Vision language model for interpretable and fine-grained detection of safety compliance in diverse workplaces
    Chen, Zhiling
    Chen, Hanning
    Imani, Mohsen
    Chen, Ruimin
    Imani, Farhad
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265