Frozen Language Model Helps ECG Zero-Shot Learning

被引:0
作者
Li, Jun [1 ]
Liu, Che [2 ,3 ]
Cheng, Sibo [3 ]
Arcucci, Rossella [2 ,3 ]
Hong, Shenda [4 ,5 ]
机构
[1] Jilin Univ, Coll Elect Sci & Engn, Changchun, Peoples R China
[2] Imperial Coll London, Dept Earth Sci & Engn, London SW7 2AZ, England
[3] Imperial Coll London, Data Sci Inst, Dept Comp, London, England
[4] Peking Univ, Natl Inst Hlth Data Sci, Beijing, Peoples R China
[5] Peking Univ, Inst Med Technol, Hlth Sci Ctr, Beijing, Peoples R China
来源
MEDICAL IMAGING WITH DEEP LEARNING, VOL 227 | 2023年 / 227卷
基金
中国国家自然科学基金;
关键词
Multimodal self-supervised learning; Zero-shot learning; Language model; ECG; Signal processing; MYOCARDIAL-INFARCTION; SIGNALS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The electrocardiogram (ECG) is one of the most commonly used non-invasive, convenient medical monitoring tools that assist in the clinical diagnosis of heart diseases. Recently, deep learning (DL) techniques, particularly self-supervised learning (SSL), have demonstrated great potential in the classification of ECG. SSL pre-training has achieved competitive performance with only a small amount of annotated data after fine-tuning. However, current SSL methods rely on the availability of annotated data and are unable to predict labels not existing in fine-tuning datasets. To address this challenge, we propose Multimodal ECG-Text Self-supervised pre-training (METS), the first work to utilize the auto-generated clinical reports to guide ECG SSL pre-training. We use a trainable ECG encoder and a frozen language model to embed paired ECG and automatically machine-generated clinical reports separately. The SSL aims to maximize the similarity between paired ECG and auto-generated report while minimize the similarity between ECG and other reports. In downstream classification tasks, METS achieves around 10% improvement in performance without using any annotated data via zero-shot classification, compared to other supervised and SSL baselines that rely on annotated data. Furthermore, METS achieves the highest recall and F1 scores on the MIT-BIH dataset, despite MIT-BIH containing different classes of ECG compared to the pre-trained dataset. The extensive experiments have demonstrated the advantages of using ECG-Text multimodal self-supervised learning in terms of generalizability, effectiveness, and efficiency.
引用
收藏
页码:402 / 415
页数:14
相关论文
共 50 条
  • [41] Class-specific synthesized dictionary model for Zero-Shot Learning
    Ji, Zhong
    Wang, Junyue
    Yu, Yunlong
    Pang, Yanwei
    Han, Jungong
    NEUROCOMPUTING, 2019, 329 : 339 - 347
  • [42] Hybrid routing transformer for zero-shot learning
    Cheng, De
    Wang, Gerong
    Wang, Bo
    Zhang, Qiang
    Han, Jungong
    Zhang, Dingwen
    PATTERN RECOGNITION, 2023, 137
  • [43] Zero-shot classification with unseen prototype learning
    Ji, Zhong
    Cui, Biying
    Yu, Yunlong
    Pang, Yanwei
    Zhang, Zhongfei
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (17) : 12307 - 12317
  • [44] Classifier and Exemplar Synthesis for Zero-Shot Learning
    Soravit Changpinyo
    Wei-Lun Chao
    Boqing Gong
    Fei Sha
    International Journal of Computer Vision, 2020, 128 : 166 - 201
  • [45] Discriminant Zero-Shot Learning with Center Loss
    Xiao-Bo Jin
    Guo-Sen Xie
    Kaizhu Huang
    Heling Cao
    Qiu-Feng Wang
    Cognitive Computation, 2019, 11 : 503 - 512
  • [46] Hierarchical Prototype Learning for Zero-Shot Recognition
    Zhang, Xingxing
    Gui, Shupeng
    Zhu, Zhenfeng
    Zhao, Yao
    Liu, Ji
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) : 1692 - 1703
  • [47] Zero-shot classification with unseen prototype learning
    Zhong Ji
    Biying Cui
    Yunlong Yu
    Yanwei Pang
    Zhongfei Zhang
    Neural Computing and Applications, 2023, 35 : 12307 - 12317
  • [48] Zero-Shot Learning for Computer Vision Applications
    Sarma, Sandipan
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 9360 - 9364
  • [49] Generative Mixup Networks for Zero-Shot Learning
    Xu, Bingrong
    Zeng, Zhigang
    Lian, Cheng
    Ding, Zhengming
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022,
  • [50] Zero-Shot Learning via Visual Abstraction
    Antol, Stanislaw
    Zitnick, C. Lawrence
    Parikh, Devi
    COMPUTER VISION - ECCV 2014, PT IV, 2014, 8692 : 401 - 416