Distilling the knowledge from large-language model for health event prediction

被引:0
作者
Ding, Sirui [1 ]
Ye, Jiancheng [2 ]
Hu, Xia [3 ]
Zou, Na [4 ]
机构
[1] Texas A&M Univ, Dept Comp Sci & Engn, College Stn, TX USA
[2] Weill Cornell Med, New York, NY USA
[3] Rice Univ, Dept Comp Sci, Houston, TX USA
[4] Univ Houston, Dept Ind Engn, Houston, TX 77004 USA
来源
SCIENTIFIC REPORTS | 2024年 / 14卷 / 01期
关键词
Health event prediction; Cardiovascular disease; Large-language model; Knowledge distillation; Multi-modal learning;
D O I
10.1038/s41598-024-75331-2
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Health event prediction is empowered by the rapid and wide application of electronic health records (EHR). In the Intensive Care Unit (ICU), precisely predicting the health related events in advance is essential for providing treatment and intervention to improve the patients outcomes. EHR is a kind of multi-modal data containing clinical text, time series, structured data, etc. Most health event prediction works focus on a single modality, e.g., text or tabular EHR. How to effectively learn from the multi-modal EHR for health event prediction remains a challenge. Inspired by the strong capability in text processing of large language model (LLM), we propose the framework CKLE for health event prediction by distilling the knowledge from LLM and learning from multi-modal EHR. There are two challenges of applying LLM in the health event prediction, the first one is most LLM can only handle text data rather than other modalities, e.g., structured data. The second challenge is the privacy issue of health applications requires the LLM to be locally deployed, which may be limited by the computational resource. CKLE solves the challenges of LLM scalability and portability in the healthcare domain by distilling the cross-modality knowledge from LLM into the health event predictive model. To fully take advantage of the strong power of LLM, the raw clinical text is refined and augmented with prompt learning. The embedding of clinical text are generated by LLM. To effectively distill the knowledge of LLM into the predictive model, we design a cross-modality knowledge distillation (KD) method. A specially designed training objective will be used for the KD process with the consideration of multiple modality and patient similarity. The KD loss function consists of two parts. The first one is cross-modality contrastive loss function, which models the correlation of different modalities from the same patient. The second one is patient similarity learning loss function to model the correlations between similar patients. The cross-modality knowledge distillation can distill the rich information in clinical text and the knowledge of LLM into the predictive model on structured EHR data. To demonstrate the effectiveness of CKLE, we evaluate CKLE on two health event prediction tasks in the field of cardiology, heart failure prediction and hypertension prediction. We select the 7125 patients from MIMIC-III dataset and split them into train/validation/test sets. We can achieve a maximum 4.48% improvement in accuracy compared to state-of-the-art predictive model designed for health event prediction. The results demonstrate CKLE can surpass the baseline prediction models significantly on both normal and limited label settings. We also conduct the case study on cardiology disease analysis in the heart failure and hypertension prediction. Through the feature importance calculation, we analyse the salient features related to the cardiology disease which corresponds to the medical domain knowledge. The superior performance and interpretability of CKLE pave a promising way to leverage the power and knowledge of LLM in the health event prediction in real-world clinical settings.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Neonatal hypertension: concerns within and beyond the neonatal intensive care unit
    Altemose, Kathleen
    Dionne, Janis M.
    [J]. CLINICAL AND EXPERIMENTAL PEDIATRICS, 2022, 65 (08) : 367 - 376
  • [2] Multimodal LLMs for Health Grounded in Individual-Specific Data
    Belyaeva, Anastasiya
    Cosentino, Justin
    Hormozdiari, Farhad
    Eswaran, Krish
    Shetty, Shravya
    Corrado, Greg
    Carroll, Andrew
    McLean, Cory Y.
    Furlotte, Nicholas A.
    [J]. MACHINE LEARNING FOR MULTIMODAL HEALTHCARE DATA, ML4MHD 2023, 2024, 14315 : 86 - 102
  • [3] Pharmacist intervention program for control of hypertension
    Chabot, I
    Moisan, J
    Grégoire, JP
    Milot, A
    [J]. ANNALS OF PHARMACOTHERAPY, 2003, 37 (09) : 1186 - 1193
  • [4] Chen T, 2020, PMLR, V119, P1597
  • [5] Choi E, 2016, ADV NEUR IN, V29
  • [6] Cohen Z. D., 2021, BERGIN GARFIELDS HDB, V7th
  • [7] High blood pressure is associated with increased risk of future fracture, but not vice versa
    Du, Xiang-Peng
    Zheng, Mei-Liang
    Yang, Xin-Chun
    Zheng, Mei-Li
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01)
  • [8] On Clinical Event Prediction in Patient Treatment Trajectory Using Longitudinal Electronic Health Records
    Duan, Huilong
    Sun, Zhoujian
    Dong, Wei
    He, Kunlun
    Huang, Zhengxing
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2020, 24 (07) : 2053 - 2063
  • [9] Characterizing the limitations of using diagnosis codes in the context of machine learning for healthcare
    Guo, Lin Lawrence
    Morse, Keith E.
    Aftandilian, Catherine
    Steinberg, Ethan
    Fries, Jason
    Posada, Jose
    Fleming, Scott Lanyon
    Lemmon, Joshua
    Jessa, Karim
    Shah, Nigam
    Sung, Lillian
    [J]. BMC MEDICAL INFORMATICS AND DECISION MAKING, 2024, 24 (01)
  • [10] Han YC, 2023, Arxiv, DOI [arXiv:2311.16483, DOI 10.48550/ARXIV.2311.16483, 10.48550/arXiv.2311.16483]