Large language models leverage external knowledge to extend clinical insight beyond language boundaries

被引:1
作者
Wu, Jiageng [1 ]
Wu, Xian [2 ]
Qiu, Zhaopeng [2 ]
Li, Minghui
Lin, Shixu [1 ]
Zhang, Yingying [2 ]
Zheng, Yefeng [2 ]
Yuan, Changzheng [1 ,3 ,5 ]
Yang, Jie [1 ,4 ,6 ,7 ]
机构
[1] Zhejiang Univ, Sch Med, Sch Publ Hlth, Hangzhou 310058, Peoples R China
[2] Tencent YouTu Lab, Jarvis Res Ctr, 1 Tianchen East Rd, Beijing 100101, Peoples R China
[3] Harvard TH Chan Sch Publ Hlth, Dept Nutr, Boston, MA 02115 USA
[4] Harvard Med Sch, Brigham & Womens Hosp, Dept Med, Div Pharmacoepidemiol & Pharmacoecon, Boston, MA 02115 USA
[5] Zhejiang Univ, Sch Publ Hlth, 866 Yuhangtang Rd, Hangzhou, Zhejiang, Peoples R China
[6] Brigham & Womens Hosp, Dept Med, 75 Francis St, Boston, MA 02115 USA
[7] Harvard Med Sch, 75 Francis St, Boston, MA 02115 USA
关键词
large language models; clinical knowledge; natural language processing; medical examination;
D O I
10.1093/jamia/ocae079
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objectives Large Language Models (LLMs) such as ChatGPT and Med-PaLM have excelled in various medical question-answering tasks. However, these English-centric models encounter challenges in non-English clinical settings, primarily due to limited clinical knowledge in respective languages, a consequence of imbalanced training corpora. We systematically evaluate LLMs in the Chinese medical context and develop a novel in-context learning framework to enhance their performance.Materials and Methods The latest China National Medical Licensing Examination (CNMLE-2022) served as the benchmark. We collected 53 medical books and 381 149 medical questions to construct the medical knowledge base and question bank. The proposed Knowledge and Few-shot Enhancement In-context Learning (KFE) framework leverages the in-context learning ability of LLMs to integrate diverse external clinical knowledge sources. We evaluated KFE with ChatGPT (GPT-3.5), GPT-4, Baichuan2-7B, Baichuan2-13B, and QWEN-72B in CNMLE-2022 and further investigated the effectiveness of different pathways for incorporating LLMs with medical knowledge from 7 distinct perspectives.Results Directly applying ChatGPT failed to qualify for the CNMLE-2022 at a score of 51. Cooperated with the KFE framework, the LLMs with varying sizes yielded consistent and significant improvements. The ChatGPT's performance surged to 70.04 and GPT-4 achieved the highest score of 82.59. This surpasses the qualification threshold (60) and exceeds the average human score of 68.70, affirming the effectiveness and robustness of the framework. It also enabled a smaller Baichuan2-13B to pass the examination, showcasing the great potential in low-resource settings.Discussion and Conclusion This study shed light on the optimal practices to enhance the capabilities of LLMs in non-English medical scenarios. By synergizing medical knowledge through in-context learning, LLMs can extend clinical insight beyond language barriers in healthcare, significantly reducing language-related disparities of LLM applications and ensuring global benefit in this field.
引用
收藏
页码:2054 / 2064
页数:11
相关论文
共 74 条
  • [1] [Anonymous], THUOCL TSINGHUA OPEN
  • [2] Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum
    Ayers, John W.
    Poliak, Adam
    Dredze, Mark
    Leas, Eric C.
    Zhu, Zechariah
    Kelley, Jessica B.
    Faix, Dennis J.
    Goodman, Aaron M.
    Longhurst, Christopher A.
    Hogarth, Michael
    Smith, Davey M.
    [J]. JAMA INTERNAL MEDICINE, 2023, 183 (06) : 589 - 596
  • [3] Bai J., 2023, ARXIV230916609, DOI DOI 10.48550/ARXIV.2309.16609
  • [4] Bang Y., 2023, P 13 INT JOINT C NAT, V1, P675
  • [5] Performance of ChatGPT on a Radiology Board-style Examination: Insights into Current Strengths and Limitations
    Bhayana, Rajesh
    Krishna, Satheesh
    Bleakney, Robert R.
    [J]. RADIOLOGY, 2023, 307 (05)
  • [6] Blevins T., 2022, P 2022 C EMPIRICAL M, P3563, DOI [DOI 10.18653/V1/2022.EMNLP-MAIN.233, 10.18653/v1/2022.emnlp-main.233]
  • [7] Brown T., 2020, ADV NEURAL INFORM PR, V33, P1877
  • [8] Chung HW, 2024, J MACH LEARN RES, V25
  • [9] Health Equity Beyond Data Health Care Worker Perceptions of Race, Ethnicity, and Language Data Collection in Electronic Health Records
    Cruz, Taylor M.
    Smith, Sheridan A.
    [J]. MEDICAL CARE, 2021, 59 (05) : 379 - 385
  • [10] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171