M3T-LM: A multi-modal multi-task learning model for jointly predicting patient length of stay and mortality

被引:0
作者
Chen, Junde [1 ]
Li, Qing [2 ]
Liu, Feng [3 ]
Wen, Yuxin [1 ]
机构
[1] Dale E. and Sarah Ann Fowler School of Engineering, Chapman University, Orange, 92866, CA
[2] Department of Industrial and Manufacturing Systems Engineering, Iowa State University, Ames, 50011, IA
[3] School of Systems and Enterprises, Stevens Institute of Technology, Hoboken, 07030, NJ
基金
美国国家科学基金会;
关键词
Data-fusion model; Deep learning; Length of stay prediction; Multi-task learning;
D O I
10.1016/j.compbiomed.2024.109237
中图分类号
学科分类号
摘要
Ensuring accurate predictions of inpatient length of stay (LoS) and mortality rates is essential for enhancing hospital service efficiency, particularly in light of the constraints posed by limited healthcare resources. Integrative analysis of heterogeneous clinic record data from different sources can hold great promise for improving the prognosis and diagnosis level of LoS and mortality. Currently, most existing studies solely focus on single data modality or tend to single-task learning, i.e., training LoS and mortality tasks separately. This limits the utilization of available multi-modal data and prevents the sharing of feature representations that could capture correlations between different tasks, ultimately hindering the model's performance. To address the challenge, this study proposes a novel Multi-Modal Multi-Task learning model, termed as M3T-LM, to integrate clinic records to predict inpatients’ LoS and mortality simultaneously. The M3T-LM framework incorporates multiple data modalities by constructing sub-models tailored to each modality. Specifically, a novel attention-embedded one-dimensional (1D) convolutional neural network (CNN) is designed to handle numerical data. For clinical notes, they are converted into sequence data, and then two long short-term memory (LSTM) networks are exploited to model on textual sequence data. A two-dimensional (2D) CNN architecture, noted as CRXMDL, is designed to extract high-level features from chest X-ray (CXR) images. Subsequently, multiple sub-models are integrated to formulate the M3T-LM to capture the correlations between patient LoS and modality prediction tasks. The efficiency of the proposed method is validated on the MIMIC-IV dataset. The proposed method attained a test MAE of 5.54 for LoS prediction and a test F1 of 0.876 for mortality prediction. The experimental results demonstrate that our approach outperforms state-of-the-art (SOTA) methods in tackling mixed regression and classification tasks. © 2024 Elsevier Ltd
引用
收藏
相关论文
共 46 条
  • [21] Multi-modal fusion and multi-task deep learning for monitoring the growth of film-mulched winter wheat
    Cheng, Zhikai
    Gu, Xiaobo
    Du, Yadan
    Wei, Chunyu
    Xu, Yang
    Zhou, Zhihui
    Li, Wenlong
    Cai, Wenjing
    PRECISION AGRICULTURE, 2024, 25 (04) : 1933 - 1957
  • [22] M3GAT: A Multi-modal, Multi-task Interactive Graph Attention Network for Conversational Sentiment Analysis and Emotion Recognition
    Zhang, Yazhou
    Jia, Ao
    Wang, Bo
    Zhang, Peng
    Zhao, Dongming
    Li, Pu
    Hou, Yuexian
    Jin, Xiaojia
    Song, Dawei
    Qin, Jing
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (01)
  • [23] Associating Multi-Modal Brain Imaging Phenotypes and Genetic Risk Factors via a Dirty Multi-Task Learning Method
    Du, Lei
    Liu, Fang
    Liu, Kefei
    Yao, Xiaohui
    Risacher, Shannon L.
    Han, Junwei
    Saykin, Andrew J.
    Shen, Li
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (11) : 3416 - 3428
  • [24] Prediction of prognostic risk factors in hepatocellular carcinoma with transarterial chemoembolization using multi-modal multi-task deep learning
    Liu, Qiu-Ping
    Xu, Xun
    Zhu, Fei-Peng
    Zhang, Yu-Dong
    Liu, Xi-Sheng
    ECLINICALMEDICINE, 2020, 23
  • [25] A collaborative multi-task model for immunohistochemical molecular sub-types of multi-modal breast cancer MRI images
    Xiang, Haozhen
    Xiong, Yuqi
    Shen, Yingwei
    Li, Jiaxin
    Liu, Deshan
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 100
  • [26] HAMMF: Hierarchical attention-based multi-task and multi-modal fusion model for computer-aided diagnosis of Alzheimer's disease
    Liu X.
    Li W.
    Miao S.
    Liu F.
    Han K.
    Bezabih T.T.
    Computers in Biology and Medicine, 2024, 176
  • [27] Neural network-based multi-task learning for inpatient flow classification and length of stay prediction
    He, Lu
    Madathil, Sreenath Chalil
    Servis, Greg
    Khasawneh, Mohammad T.
    APPLIED SOFT COMPUTING, 2021, 108
  • [28] Robust texture analysis of multi-modal images using Local Structure Preserving Ranklet and multi-task learning for breast tumor diagnosis
    Xi, Xiaoming
    Xu, Hui
    Shi, Hao
    Zhang, Chunyun
    Ding, Hong Yu
    Zhang, Guang
    Tang, Yuchun
    Yin, Yilong
    NEUROCOMPUTING, 2017, 259 : 210 - 218
  • [29] M3IL: Multi-Modal Meta-Imitation Learning
    Zhang X.
    Matsushima T.
    Matsuo Y.
    Iwasawa Y.
    Transactions of the Japanese Society for Artificial Intelligence, 2023, 38 (02)
  • [30] A Photovoltaic Power Predicting Model Using the Differential Evolution Algorithm and Multi-Task Learning
    Pang, Senshen
    Liu, Jiefeng
    Zhang, Zhenhao
    Fan, Xianhao
    Zhang, Yiyi
    Zhang, Dongdong
    Hwang, Goh Hui
    FRONTIERS IN MATERIALS, 2022, 9