Collaborative networks of transformers and convolutional neural networks are powerful and versatile learners for accurate 3D medical image segmentation

被引:6
作者
Chen, Yong [1 ]
Lu, Xuesong [1 ]
Xie, Qinlan [1 ]
机构
[1] South Cent Minzu Univ, Sch Biomed Engn, Wuhan 430074, Hubei, Peoples R China
关键词
Convolutional neural networks; Transformers; Interlaced collaboration; Versatile models; 3D medical image segmentation;
D O I
10.1016/j.compbiomed.2023.107228
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Integrating transformers and convolutional neural networks represents a crucial and cutting-edge approach for tackling medical image segmentation problems. Nonetheless, the existing hybrid methods fail to fully leverage the strengths of both operators. During the Patch Embedding, the patch projection method ignores the two-dimensional structure and local spatial information within each patch, while the fixed patch size cannot capture features with rich representation effectively. Moreover, the calculation of self-attention results in attention diffusion, hindering the provision of precise details to the decoder while maintaining feature consistency. Lastly, none of the existing methods establish an efficient multi-scale modeling concept. To address these issues, we design the Collaborative Networks of Transformers and Convolutional neural networks (TC-CoNet), which is generally used for accurate 3D medical image segmentation. First, we elaborately design precise patch embedding to generate 3D features with accurate spatial position information, laying a solid foundation for subsequent learning. The encoder-decoder backbone network is then constructed by TC-CoNet in an interlaced combination to properly incorporate long-range dependencies and hierarchical object concepts at various scales. Furthermore, we employ the constricted attention bridge to constrict attention to local features, allowing us to accurately guide the recovery of detailed information while maintaining feature consistency. Finally, atrous spatial pyramid pooling is applied to high-level feature map to establish the concept of multi-scale objects. On five challenging datasets, including Synapse, ACDC, brain tumor segmentation, cardiac left atrium segmentation, and lung tumor segmentation, the extensive experiments demonstrate that TC-CoNet outperforms state-of-the-art approaches in terms of superiority, migration, and strong generalization. These illustrate in full the efficacy of the proposed transformers and convolutional neural networks combination for medical image segmentation. Our code is freely available at: https://github.com/YongChen-Exact/TC-CoNet.
引用
收藏
页数:14
相关论文
共 69 条
  • [1] The Medical Segmentation Decathlon
    Antonelli, Michela
    Reinke, Annika
    Bakas, Spyridon
    Farahani, Keyvan
    Kopp-Schneider, Annette
    Landman, Bennett A.
    Litjens, Geert
    Menze, Bjoern
    Ronneberger, Olaf
    Summers, Ronald M.
    van Ginneken, Bram
    Bilello, Michel
    Bilic, Patrick
    Christ, Patrick F.
    Do, Richard K. G.
    Gollub, Marc J.
    Heckers, Stephan H.
    Huisman, Henkjan
    Jarnagin, William R.
    McHugo, Maureen K.
    Napel, Sandy
    Pernicka, Jennifer S. Golia
    Rhode, Kawal
    Tobon-Gomez, Catalina
    Vorontsov, Eugene
    Meakin, James A.
    Ourselin, Sebastien
    Wiesenfarth, Manuel
    Arbelaez, Pablo
    Bae, Byeonguk
    Chen, Sihong
    Daza, Laura
    Feng, Jianjiang
    He, Baochun
    Isensee, Fabian
    Ji, Yuanfeng
    Jia, Fucang
    Kim, Ildoo
    Maier-Hein, Klaus
    Merhof, Dorit
    Pai, Akshay
    Park, Beomhee
    Perslev, Mathias
    Rezaiifar, Ramin
    Rippel, Oliver
    Sarasua, Ignacio
    Shen, Wei
    Son, Jaemin
    Wachinger, Christian
    Wang, Liansheng
    [J]. NATURE COMMUNICATIONS, 2022, 13 (01)
  • [2] Azad R, 2023, Arxiv, DOI [arXiv:2301.10847, DOI 10.48550/ARXIV.2301.10847]
  • [3] Azad R, 2022, Arxiv, DOI [arXiv:2212.13504, DOI 10.48550/ARXIV.2212.13504]
  • [4] TransDeepLab: Convolution-Free Transformer-Based DeepLab v3+for Medical Image Segmentation
    Azad, Reza
    Heidari, Moein
    Shariatnia, Moein
    Aghdam, Ehsan Khodapanah
    Karimijafarbigloo, Sanaz
    Adeli, Ehsan
    Merhof, Dorit
    [J]. PREDICTIVE INTELLIGENCE IN MEDICINE (PRIME 2022), 2022, 13564 : 91 - 102
  • [5] TransNorm: Transformer Provides a Strong Spatial Normalization Mechanism for a Deep Segmentation Model
    Azad, Reza
    Al-Antary, Mohammad T.
    Heidari, Moein
    Merhof, Dorit
    [J]. IEEE ACCESS, 2022, 10 : 108205 - 108215
  • [6] Hybrid-scale contextual fusion network for medical image segmentation
    Bao, Hua
    Zhu, Yuqing
    Li, Qing
    [J]. COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 152
  • [7] An Exploration of 2D and 3D Deep Learning Techniques for Cardiac MR Image Segmentation
    Baumgartner, Christian F.
    Koch, Lisa M.
    Pollefeys, Marc
    Konukoglu, Ender
    [J]. STATISTICAL ATLASES AND COMPUTATIONAL MODELS OF THE HEART: ACDC AND MMWHS CHALLENGES, 2018, 10663 : 111 - 119
  • [8] Deep Learning Techniques for Automatic MRI Cardiac Multi-Structures Segmentation and Diagnosis: Is the Problem Solved?
    Bernard, Olivier
    Lalande, Alain
    Zotti, Clement
    Cervenansky, Frederick
    Yang, Xin
    Heng, Pheng-Ann
    Cetin, Irem
    Lekadir, Karim
    Camara, Oscar
    Gonzalez Ballester, Miguel Angel
    Sanroma, Gerard
    Napel, Sandy
    Petersen, Steffen
    Tziritas, Georgios
    Grinias, Elias
    Khened, Mahendra
    Kollerathu, Varghese Alex
    Krishnamurthi, Ganapathy
    Rohe, Marc-Michel
    Pennec, Xavier
    Sermesant, Maxime
    Isensee, Fabian
    Jaeger, Paul
    Maier-Hein, Klaus H.
    Full, Peter M.
    Wolf, Ivo
    Engelhardt, Sandy
    Baumgartner, Christian F.
    Koch, Lisa M.
    Wolterink, Jelmer M.
    Isgum, Ivana
    Jang, Yeonggul
    Hong, Yoonmi
    Patravali, Jay
    Jain, Shubham
    Humbert, Olivier
    Jodoin, Pierre-Marc
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2018, 37 (11) : 2514 - 2525
  • [9] Cao Hu, 2023, Computer Vision - ECCV 2022 Workshops: Proceedings. Lecture Notes in Computer Science (13803), P205, DOI 10.1007/978-3-031-25066-8_9
  • [10] Chen BZ, 2022, Arxiv, DOI arXiv:2107.05274