Frozen Language Model Helps ECG Zero-Shot Learning

被引:0
作者
Li, Jun [1 ]
Liu, Che [2 ,3 ]
Cheng, Sibo [3 ]
Arcucci, Rossella [2 ,3 ]
Hong, Shenda [4 ,5 ]
机构
[1] Jilin Univ, Coll Elect Sci & Engn, Changchun, Peoples R China
[2] Imperial Coll London, Dept Earth Sci & Engn, London SW7 2AZ, England
[3] Imperial Coll London, Data Sci Inst, Dept Comp, London, England
[4] Peking Univ, Natl Inst Hlth Data Sci, Beijing, Peoples R China
[5] Peking Univ, Inst Med Technol, Hlth Sci Ctr, Beijing, Peoples R China
来源
MEDICAL IMAGING WITH DEEP LEARNING, VOL 227 | 2023年 / 227卷
基金
中国国家自然科学基金;
关键词
Multimodal self-supervised learning; Zero-shot learning; Language model; ECG; Signal processing; MYOCARDIAL-INFARCTION; SIGNALS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The electrocardiogram (ECG) is one of the most commonly used non-invasive, convenient medical monitoring tools that assist in the clinical diagnosis of heart diseases. Recently, deep learning (DL) techniques, particularly self-supervised learning (SSL), have demonstrated great potential in the classification of ECG. SSL pre-training has achieved competitive performance with only a small amount of annotated data after fine-tuning. However, current SSL methods rely on the availability of annotated data and are unable to predict labels not existing in fine-tuning datasets. To address this challenge, we propose Multimodal ECG-Text Self-supervised pre-training (METS), the first work to utilize the auto-generated clinical reports to guide ECG SSL pre-training. We use a trainable ECG encoder and a frozen language model to embed paired ECG and automatically machine-generated clinical reports separately. The SSL aims to maximize the similarity between paired ECG and auto-generated report while minimize the similarity between ECG and other reports. In downstream classification tasks, METS achieves around 10% improvement in performance without using any annotated data via zero-shot classification, compared to other supervised and SSL baselines that rely on annotated data. Furthermore, METS achieves the highest recall and F1 scores on the MIT-BIH dataset, despite MIT-BIH containing different classes of ECG compared to the pre-trained dataset. The extensive experiments have demonstrated the advantages of using ECG-Text multimodal self-supervised learning in terms of generalizability, effectiveness, and efficiency.
引用
收藏
页码:402 / 415
页数:14
相关论文
共 50 条
  • [21] A review on multimodal zero-shot learning
    Cao, Weipeng
    Wu, Yuhao
    Sun, Yixuan
    Zhang, Haigang
    Ren, Jin
    Gu, Dujuan
    Wang, Xingkai
    WILEY INTERDISCIPLINARY REVIEWS-DATA MINING AND KNOWLEDGE DISCOVERY, 2023, 13 (02)
  • [22] Attribute subspaces for zero-shot learning
    Zhou, Lei
    Liu, Yang
    Bai, Xiao
    Li, Na
    Yu, Xiaohan
    Zhou, Jun
    Hancock, Edwin R.
    PATTERN RECOGNITION, 2023, 144
  • [23] LVQ Treatment for Zero-Shot Learning
    Ismailoglu, Firat
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2023, 31 (01) : 216 - 237
  • [24] Towards Zero-Shot Sign Language Recognition
    Bilge, Yunus Can
    Cinbis, Ramazan Gokberk
    Ikizler-Cinbis, Nazli
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (01) : 1217 - 1232
  • [25] Learning MLatent Representations for Generalized Zero-Shot Learning
    Ye, Yalan
    Pan, Tongjie
    Luo, Tonghoujun
    Li, Jingjing
    Shen, Heng Tao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 2252 - 2265
  • [26] A Zero-Shot Learning Approach for Blockage Detection and Identification Based on the Stacking Ensemble Model
    Li, Chaoqun
    Feng, Zao
    Jiang, Mingkai
    Wang, Zhenglang
    SENSORS, 2024, 24 (17)
  • [27] Semantic-aligned reinforced attention model for zero-shot learning
    Yang, Zaiquan
    Zhang, Yuqi
    Du, Yuxin
    Tong, Chao
    IMAGE AND VISION COMPUTING, 2022, 128
  • [28] Harnessing large language models' zero-shot and few-shot learning capabilities for regulatory research
    Meshkin, Hamed
    Zirkle, Joel
    Arabidarrehdor, Ghazal
    Chaturbedi, Anik
    Chakravartula, Shilpa
    Mann, John
    Thrasher, Bradlee
    Li, Zhihua
    BRIEFINGS IN BIOINFORMATICS, 2024, 25 (05)
  • [29] Structure Fusion and Propagation for Zero-Shot Learning
    Lin, Guangfeng
    Chen, Yajun
    Zhao, Fan
    PATTERN RECOGNITION AND COMPUTER VISION, PT III, 2018, 11258 : 465 - 477
  • [30] Integrative zero-shot learning for fruit recognition
    Tran-Anh, Dat
    Huu, Quynh Nguyen
    Bui-Quoc, Bao
    Hoang, Ngan Dao
    Quoc, Tao Ngo
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (29) : 73191 - 73213