Multi-modal learning for inpatient length of stay prediction

被引:7
作者
Chen, Junde [1 ]
Wen, Yuxin [1 ]
Pokojovy, Michael [2 ]
Tseng, Tzu-Liang [3 ]
McCaffrey, Peter [4 ]
Vo, Alexander [4 ]
Walser, Eric [4 ]
Moen, Scott [4 ]
机构
[1] Chapman Univ, Dale E & Sarah Ann Fowler Sch Engn, Orange, CA 92866 USA
[2] Old Dominion Univ, Dept Math & Stat, Norfolk, VA 23529 USA
[3] Univ Texas El Paso, Dept Ind Mfg & Syst Engn, El Paso, TX 79968 USA
[4] Univ Texas Med Branch, Galveston, TX 77550 USA
基金
美国国家科学基金会;
关键词
Chest X-ray images; Data -fusion model; Length of stay prediction; Multi -modal learning; HOSPITAL MORTALITY;
D O I
10.1016/j.compbiomed.2024.108121
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Predicting inpatient length of stay (LoS) is important for hospitals aiming to improve service efficiency and enhance management capabilities. Patient medical records are strongly associated with LoS. However, due to diverse modalities, heterogeneity, and complexity of data, it becomes challenging to effectively leverage these heterogeneous data to put forth a predictive model that can accurately predict LoS. To address the challenge, this study aims to establish a novel data-fusion model, termed as DF-Mdl, to integrate heterogeneous clinical data for predicting the LoS of inpatients between hospital discharge and admission. Multi-modal data such as demographic data, clinical notes, laboratory test results, and medical images are utilized in our proposed methodology with individual "basic" sub-models separately applied to each different data modality. Specifically, a convolutional neural network (CNN) model, which we termed CRXMDL, is designed for chest X-ray (CXR) image data, two long short-term memory networks are used to extract features from long text data, and a novel attention-embedded 1D convolutional neural network is developed to extract useful information from numerical data. Finally, these basic models are integrated to form a new data-fusion model (DF-Mdl) for inpatient LoS prediction. The proposed method attains the best R2 and EVAR values of 0.6039 and 0.6042 among competitors for the LoS prediction on the Medical Information Mart for Intensive Care (MIMIC)-IV test dataset. Empirical evidence suggests better performance compared with other state-of-the-art (SOTA) methods, which demonstrates the effectiveness and feasibility of the proposed approach.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Multi-modal learning for affective content analysis in movies
    Yun Yi
    Hanli Wang
    Multimedia Tools and Applications, 2019, 78 : 13331 - 13350
  • [22] Knowledge Graphs for Multi-modal Learning: Survey and Perspective
    Chen, Zhuo
    Zhang, Yichi
    Fang, Yin
    Geng, Yuxia
    Guo, Lingbing
    Chen, Jiaoyan
    Liu, Xiaoze
    Pan, Jeff Z.
    Zhang, Ningyu
    Chen, Huajun
    Zhang, Wen
    INFORMATION FUSION, 2025, 121
  • [23] Multi-modal learning and its application for biomedical data
    Liu, Jin
    Zhang, Yu-Dong
    Cai, Hongming
    FRONTIERS IN MEDICINE, 2024, 10
  • [24] Bayesian mixture variational autoencoders for multi-modal learning
    Liao, Keng-Te
    Huang, Bo-Wei
    Yang, Chih-Chun
    Lin, Shou-De
    MACHINE LEARNING, 2022, 111 (12) : 4329 - 4357
  • [25] Bayesian mixture variational autoencoders for multi-modal learning
    Keng-Te Liao
    Bo-Wei Huang
    Chih-Chun Yang
    Shou-De Lin
    Machine Learning, 2022, 111 : 4329 - 4357
  • [26] Inpatient Length of Stay and Mortality Prediction Utilizing Clinical Time Series Data
    Chen, Junde
    Li, Mason
    Milosevich, Miles
    Le, Tiffany
    Bahsoun, Andrew
    Wen, Yuxin
    IEEE ACCESS, 2025, 13 : 74720 - 74734
  • [27] Multi-modal Machine Learning Model for Interpretable Malware Classification
    Lisa, Fahmida Tasnim
    Islam, Sheikh Rabiul
    Kumar, Neha Mohan
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, PT III, XAI 2024, 2024, 2155 : 334 - 349
  • [28] Graph Embedding Contrastive Multi-Modal Representation Learning for Clustering
    Xia, Wei
    Wang, Tianxiu
    Gao, Quanxue
    Yang, Ming
    Gao, Xinbo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 1170 - 1183
  • [29] A multi-modal sarcasm detection model based on cue learning
    Lu, Ming
    Dong, Zhiqiang
    Guo, Ziming
    Zhang, Xiaoming
    Lu, Xinxi
    Wang, Tianbo
    Zhang, Litian
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [30] Correspondence Learning for Deep Multi-Modal Recognition and Fraud Detection
    Park, Jongchan
    Kim, Min-Hyun
    Choi, Dong-Geol
    ELECTRONICS, 2021, 10 (07)