DCCAT: Dual-Coordinate Cross-Attention Transformer for thrombus segmentation on coronary OCT

被引:2
作者
Chu, Miao [1 ,2 ,3 ]
De Maria, Giovanni Luigi [2 ,3 ]
Dai, Ruobing [1 ]
Benenati, Stefano [2 ,3 ,5 ]
Yu, Wei [1 ]
Zhong, Jiaxin [1 ,6 ]
Kotronias, Rafail [2 ,3 ,4 ]
Walsh, Jason [2 ,3 ,4 ]
Andreaggi, Stefano [2 ,7 ]
Zuccarelli, Vittorio [2 ]
Chai, Jason [2 ,3 ]
Channon, Keith [2 ,3 ,4 ]
Banning, Adrian [2 ,3 ,4 ]
Tu, Shengxian [1 ,3 ]
机构
[1] Shanghai Jiao Tong Univ, Biomed Instrument Inst, Sch Biomed Engn, Shanghai, Peoples R China
[2] Oxford Univ Hosp NHS Trust, Oxford Heart Ctr, Oxford, England
[3] Univ Oxford, Radcliffe Dept Med, Div Cardiovasc Med, Oxford, England
[4] Oxford Biomed Res Ctr, Natl Inst Hlth Res, Oxford, England
[5] Univ Genoa, Genoa, Italy
[6] Fujian Med Univ, Union Hosp, Dept Cardiol, Fuzhou, Fujian, Peoples R China
[7] Univ Verona, Dept Med, Div Cardiol, Verona, Italy
基金
中国国家自然科学基金;
关键词
Acute coronary syndromes; Optical coherence tomography; Thrombus segmentation; Cross-attention; OPTICAL COHERENCE TOMOGRAPHY; PLAQUE EROSION; NEURAL-NETWORK; DIAGNOSIS;
D O I
10.1016/j.media.2024.103265
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Acute coronary syndromes (ACS) are one of the leading causes of mortality worldwide, with atherosclerotic plaque rupture and subsequent thrombus formation as the main underlying substrate. Thrombus burden evaluation is important for tailoring treatment therapy and predicting prognosis. Coronary optical coherence tomography (OCT) enables in-vivo visualization of thrombus that cannot otherwise be achieved by other image modalities. However, automatic quantification of thrombus on OCT has not been implemented. The main challenges are due to the variation in location, size and irregularities of thrombus in addition to the small data set. In this paper, we propose a novel dual-coordinate cross-attention transformer network, termed DCCAT, to overcome the above challenges and achieve the first automatic segmentation of thrombus on OCT. Imaging features from both Cartesian and polar coordinates are encoded and fused based on long-range correspondence via multi-head cross-attention mechanism. The dual-coordinate cross-attention block is hierarchically stacked amid convolutional layers at multiple levels, allowing comprehensive feature enhancement. The model was developed based on 5,649 OCT frames from 339 patients and tested using independent external OCT data from 548 frames of 52 patients. DCCAT achieved Dice similarity score (DSC) of 0.706 in segmenting thrombus, which is significantly higher than the CNN-based (0.656) and Transformer-based (0.584) models. We prove that the additional input of polar image not only leverages discriminative features from another coordinate but also improves model robustness for geometrical transformation.Experiment results show that DCCAT achieves competitive performance with only 10% of the total data, highlighting its data efficiency. The proposed dual- coordinate cross-attention design can be easily integrated into other developed Transformer models to boost performance.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] LGCANet:Local geometry-aware cross-attention networks for point cloud semantic segmentation
    Luo, Liguo
    Lu, Jian
    Chen, Xiaogai
    Zhang, Kaibing
    Zhou, Jian
    MEASUREMENT, 2025, 248
  • [32] ABC-Trans: a novel adaptive border-augmented cross-attention transformer for object detection
    Qianjun Zhang
    Pan Wang
    Zihao Wu
    Binhong Yang
    Jin Yuan
    Multimedia Tools and Applications, 2025, 84 (16) : 15671 - 15688
  • [33] A cross-attention integrated shifted window transformer for remote sensing image scene recognition with limited data
    Li, Kaiyuan
    Xue, Yong
    Zhao, Jiaqi
    Li, Honghao
    Zhang, Sheng
    JOURNAL OF APPLIED REMOTE SENSING, 2024, 18 (03)
  • [34] MM-UNet: A novel cross-attention mechanism between modules and scales for brain tumor segmentation
    Lin, Chih-Wei
    Chen, Zhongsheng
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [35] Pixel Representation Augmented through Cross-Attention for High-Resolution Remote Sensing Imagery Segmentation
    Luo, Yiyun
    Wang, Jinnian
    Yang, Xiankun
    Yu, Zhenyu
    Tan, Zixuan
    REMOTE SENSING, 2022, 14 (21)
  • [36] CerviFormer: A pap smear-based cervical cancer classification method using cross-attention and latent transformer
    Deo, Bhaswati Singha
    Pal, Mayukha
    Panigrahi, Prasanta K.
    Pradhan, Asima
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2024, 34 (02)
  • [37] Twins transformer: rolling bearing fault diagnosis based on cross-attention fusion of time and frequency domain features
    Gao, Zhikang
    Wang, Yanxue
    Li, Xinming
    Yao, Jiachi
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (09)
  • [38] U-Net Transformer: Self and Cross Attention for Medical Image Segmentation
    Petit, Olivier
    Thome, Nicolas
    Rambour, Clement
    Themyr, Loic
    Collins, Toby
    Soler, Luc
    MACHINE LEARNING IN MEDICAL IMAGING, MLMI 2021, 2021, 12966 : 267 - 276
  • [39] A cross-attention swin transformer network for EEG-based subject-independent cognitive load assessment
    Li, Zhongrui
    Zhang, Rongkai
    Tong, Li
    Zeng, Ying
    Gao, Yuanlong
    Yang, Kai
    Yan, Bin
    COGNITIVE NEURODYNAMICS, 2024, : 3805 - 3819
  • [40] Unsupervised Domain Adaptive Dose Prediction Via Cross-Attention Transformer and Target-Specific Knowledge Preservation
    Cui, Jiaqi
    Xiao, Jianghong
    Hou, Yun
    Wu, Xi
    Zhou, Jiliu
    Peng, Xingchen
    Wang, Yan
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2023, 33 (11)