Multi-modal learning for inpatient length of stay prediction

被引:7
作者
Chen, Junde [1 ]
Wen, Yuxin [1 ]
Pokojovy, Michael [2 ]
Tseng, Tzu-Liang [3 ]
McCaffrey, Peter [4 ]
Vo, Alexander [4 ]
Walser, Eric [4 ]
Moen, Scott [4 ]
机构
[1] Chapman Univ, Dale E & Sarah Ann Fowler Sch Engn, Orange, CA 92866 USA
[2] Old Dominion Univ, Dept Math & Stat, Norfolk, VA 23529 USA
[3] Univ Texas El Paso, Dept Ind Mfg & Syst Engn, El Paso, TX 79968 USA
[4] Univ Texas Med Branch, Galveston, TX 77550 USA
基金
美国国家科学基金会;
关键词
Chest X-ray images; Data -fusion model; Length of stay prediction; Multi -modal learning; HOSPITAL MORTALITY;
D O I
10.1016/j.compbiomed.2024.108121
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Predicting inpatient length of stay (LoS) is important for hospitals aiming to improve service efficiency and enhance management capabilities. Patient medical records are strongly associated with LoS. However, due to diverse modalities, heterogeneity, and complexity of data, it becomes challenging to effectively leverage these heterogeneous data to put forth a predictive model that can accurately predict LoS. To address the challenge, this study aims to establish a novel data-fusion model, termed as DF-Mdl, to integrate heterogeneous clinical data for predicting the LoS of inpatients between hospital discharge and admission. Multi-modal data such as demographic data, clinical notes, laboratory test results, and medical images are utilized in our proposed methodology with individual "basic" sub-models separately applied to each different data modality. Specifically, a convolutional neural network (CNN) model, which we termed CRXMDL, is designed for chest X-ray (CXR) image data, two long short-term memory networks are used to extract features from long text data, and a novel attention-embedded 1D convolutional neural network is developed to extract useful information from numerical data. Finally, these basic models are integrated to form a new data-fusion model (DF-Mdl) for inpatient LoS prediction. The proposed method attains the best R2 and EVAR values of 0.6039 and 0.6042 among competitors for the LoS prediction on the Medical Information Mart for Intensive Care (MIMIC)-IV test dataset. Empirical evidence suggests better performance compared with other state-of-the-art (SOTA) methods, which demonstrates the effectiveness and feasibility of the proposed approach.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] A Multi-Modal Active Learning Experience for Teaching Social Categorization
    Schwarzmueller, April
    TEACHING OF PSYCHOLOGY, 2011, 38 (03) : 158 - 161
  • [32] SPARSE TIKHONOV-REGULARIZED HASHING FOR MULTI-MODAL LEARNING
    Tian, Lei
    Hong, Xiaopeng
    Fan, Chunxiao
    Ming, Yue
    Pietikainen, Matti
    Zhao, Guoying
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 3793 - 3797
  • [33] Learning multi-modal recurrent neural networks with target propagation
    Manchev, Nikolay
    Spratling, Michael
    COMPUTATIONAL INTELLIGENCE, 2024, 40 (04)
  • [34] Learning Multi-Modal Nonlinear Embeddings: Performance Bounds and an Algorithm
    Kaya, Semih
    Vural, Elif
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 4384 - 4394
  • [35] Multi-modal deep learning for automated assembly of periapical radiographs
    Pfaender, L.
    Schneider, L.
    Buettner, M.
    Krois, J.
    Meyer-Lueckel, H.
    Schwendicke, F.
    JOURNAL OF DENTISTRY, 2023, 135
  • [36] Learnable Cross-modal Knowledge Distillation for Multi-modal Learning with Missing Modality
    Wang, Hu
    Ma, Congbo
    Zhang, Jianpeng
    Zhang, Yuan
    Avery, Jodie
    Hull, Louise
    Carneiro, Gustavo
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 216 - 226
  • [37] SusTriage: Sustainable Bug Triage with Multi-modal Ensemble Learning
    Zhang, Wen
    Zhao, Jiangpeng
    Wang, Song
    PROCEEDINGS OF 2021 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY WORKSHOPS AND SPECIAL SESSIONS: (WI-IAT WORKSHOP/SPECIAL SESSION 2021), 2021, : 441 - 448
  • [38] A Theoretical Analysis of Multi-Modal Representation Learning with Regular Functions
    Vural, Elif
    2020 28TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2020,
  • [39] Multi-Task and Multi-Modal Learning for RGB Dynamic Gesture Recognition
    Fan, Dinghao
    Lu, Hengjie
    Xu, Shugong
    Cao, Shan
    IEEE SENSORS JOURNAL, 2021, 21 (23) : 27026 - 27036
  • [40] MGDR: Multi-modal Graph Disentangled Representation for Brain Disease Prediction
    Jiang, Bo
    Li, Yapeng
    Wan, Xixi
    Chen, Yuan
    Tu, Zhengzheng
    Zhao, Yumiao
    Tang, Jin
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT II, 2024, 15002 : 302 - 312