M3T-LM: A multi-modal multi-task learning model for jointly predicting patient length of stay and mortality

被引:0
作者
Chen, Junde [1 ]
Li, Qing [2 ]
Liu, Feng [3 ]
Wen, Yuxin [1 ]
机构
[1] Dale E. and Sarah Ann Fowler School of Engineering, Chapman University, Orange, 92866, CA
[2] Department of Industrial and Manufacturing Systems Engineering, Iowa State University, Ames, 50011, IA
[3] School of Systems and Enterprises, Stevens Institute of Technology, Hoboken, 07030, NJ
基金
美国国家科学基金会;
关键词
Data-fusion model; Deep learning; Length of stay prediction; Multi-task learning;
D O I
10.1016/j.compbiomed.2024.109237
中图分类号
学科分类号
摘要
Ensuring accurate predictions of inpatient length of stay (LoS) and mortality rates is essential for enhancing hospital service efficiency, particularly in light of the constraints posed by limited healthcare resources. Integrative analysis of heterogeneous clinic record data from different sources can hold great promise for improving the prognosis and diagnosis level of LoS and mortality. Currently, most existing studies solely focus on single data modality or tend to single-task learning, i.e., training LoS and mortality tasks separately. This limits the utilization of available multi-modal data and prevents the sharing of feature representations that could capture correlations between different tasks, ultimately hindering the model's performance. To address the challenge, this study proposes a novel Multi-Modal Multi-Task learning model, termed as M3T-LM, to integrate clinic records to predict inpatients’ LoS and mortality simultaneously. The M3T-LM framework incorporates multiple data modalities by constructing sub-models tailored to each modality. Specifically, a novel attention-embedded one-dimensional (1D) convolutional neural network (CNN) is designed to handle numerical data. For clinical notes, they are converted into sequence data, and then two long short-term memory (LSTM) networks are exploited to model on textual sequence data. A two-dimensional (2D) CNN architecture, noted as CRXMDL, is designed to extract high-level features from chest X-ray (CXR) images. Subsequently, multiple sub-models are integrated to formulate the M3T-LM to capture the correlations between patient LoS and modality prediction tasks. The efficiency of the proposed method is validated on the MIMIC-IV dataset. The proposed method attained a test MAE of 5.54 for LoS prediction and a test F1 of 0.876 for mortality prediction. The experimental results demonstrate that our approach outperforms state-of-the-art (SOTA) methods in tackling mixed regression and classification tasks. © 2024 Elsevier Ltd
引用
收藏
相关论文
共 46 条
  • [11] Multi-Modal Meta Multi-Task Learning For Social Media Rumor Detection
    Poornima, R.
    Nagavarapu, Sateesh
    Navya, Soleti
    Katkoori, Arun Kumar
    Mohsen, Karrar Shareef
    Saikumar, K.
    2024 2ND WORLD CONFERENCE ON COMMUNICATION & COMPUTING, WCONF 2024, 2024,
  • [12] Adversarial Multi-Task Learning for Mandarin Prosodic Boundary Prediction With Multi-Modal Embeddings
    Yi, Jiangyan
    Tao, Jianhua
    Fu, Ruibo
    Wang, Tao
    Zhang, Chu Yuan
    Wang, Chenglong
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 2963 - 2973
  • [13] Align vision-language semantics by multi-task learning for multi-modal summarization
    Cui C.
    Liang X.
    Wu S.
    Li Z.
    Neural Computing and Applications, 2024, 36 (25) : 15653 - 15666
  • [14] MBFusion: Multi-modal balanced fusion and multi-task learning for cancer diagnosis and prognosis
    Zhang, Ziye
    Yin, Wendong
    Wang, Shijin
    Zheng, Xiaorou
    Dong, Shoubin
    Computers in Biology and Medicine, 2024, 181
  • [15] Cloud Type Classification Using Multi-modal Information Based on Multi-task Learning
    Zhang, Yaxiu
    Xie, Jiazu
    He, Di
    Dong, Qing
    Zhang, Jiafeng
    Zhang, Zhong
    Liu, Shuang
    COMMUNICATIONS, SIGNAL PROCESSING, AND SYSTEMS, VOL. 1, 2022, 878 : 119 - 125
  • [16] Fake News Detection in Social Media based on Multi-Modal Multi-Task Learning
    Cui, Xinyu
    Li, Yang
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (07) : 912 - 918
  • [17] Multi-Task Federated Split Learning Across Multi-Modal Data with Privacy Preservation
    Dong, Yipeng
    Luo, Wei
    Wang, Xiangyang
    Zhang, Lei
    Xu, Lin
    Zhou, Zehao
    Wang, Lulu
    SENSORS, 2025, 25 (01)
  • [18] A multi-modal fusion framework based on multi-task correlation learning for cancer prognosis prediction
    Tan, Kaiwen
    Huang, Weixian
    Liu, Xiaofeng
    Hu, Jinlong
    Dong, Shoubin
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2022, 126
  • [19] Multi-spatial urban function modeling: A multi-modal deep network approach for transfer and multi-task learning
    Gong, Zhaoya
    Wang, Chenglong
    Liu, Bin
    Li, Binbo
    Tu, Wei
    Chen, Yuting
    Deng, Zhicheng
    Zhao, Pengjun
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2025, 136
  • [20] MM-DAG: Multi-task DAG Learning for Multi-modal Data - with Application for Traffic Congestion Analysis
    Lan, Tian
    Li, Ziyue
    Li, Zhishuai
    Bai, Lei
    Li, Man
    Tsung, Fugee
    Ketter, Wolfgang
    Zhao, Rui
    Zhang, Chen
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 1188 - 1199