Convolution-Free Medical Image Segmentation Using Transformers

被引:90
作者
Karimi, Davood [1 ]
Vasylechko, Serge Didenko
Gholipour, Ali
机构
[1] Boston Childrens Hosp, Dept Radiol, Computat Radiol Lab CRL, Boston, MA 02115 USA
来源
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT I | 2021年 / 12901卷
基金
美国国家卫生研究院;
关键词
Segmentation; Deep learning; Transformers; Attention;
D O I
10.1007/978-3-030-87193-2_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Like other applications in computer vision, medical image segmentation and his email address have been most successfully addressed using deep learning models that rely on the convolution operation as their main building block. Convolutions enjoy important properties such as sparse interactions, weight sharing, and translation equivariance. These properties give convolutional neural networks (CNNs) a strong and useful inductive bias for vision tasks. However, the convolution operation also has important shortcomings: it performs a fixed operation on every test image regardless of the content and it cannot efficiently model long-range interactions. In this work we show that a network based on self-attention between neighboring patches and without any convolution operations can achieve better results. Given a 3D image block, our network divides it into n(3) 3D patches, where n = 3 or 5 and computes a 1D embedding for each patch. The network predicts the segmentation map for the center patch of the block based on the self-attention between these patch embeddings. We show that the proposed model can achieve higher segmentation accuracies than a state of the art CNN. For scenarios with very few labeled images, we propose methods for pre-training the network on large corpora of unlabeled images. Our experiments show that with pre-training the advantage of our proposed network over CNNs can be significant when labeled training data is small.
引用
收藏
页码:78 / 88
页数:11
相关论文
共 34 条
[1]  
[Anonymous], 2018, I S BIOMED IMAGING
[2]  
[Anonymous], Food Additives Contaminants
[3]   Deep semantic segmentation of natural and medical images: a review [J].
Asgari Taghanaki, Saeid ;
Abhishek, Kumar ;
Cohen, Joseph Paul ;
Cohen-Adad, Julien ;
Hamarneh, Ghassan .
ARTIFICIAL INTELLIGENCE REVIEW, 2021, 54 (01) :137-178
[4]   Recurrent Neural Networks for Aortic Image Sequence Segmentation with Sparse Annotations [J].
Bai, Wenjia ;
Suzuki, Hideaki ;
Qin, Chen ;
Tarroni, Giacomo ;
Oktay, Ozan ;
Matthews, Paul M. ;
Rueckert, Daniel .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2018, PT IV, 2018, 11073 :586-594
[5]  
Bakas S., 2018, arXiv
[6]   Automated processing pipeline for neonatal diffusion MRI in the developing Human Connectome Project [J].
Bastiani, Matteo ;
Andersson, Jesper L. R. ;
Cordero-Grande, Lucilio ;
Murgasova, Maria ;
Hutter, Jana ;
Price, Anthony N. ;
Makropoulos, Antonios ;
Fitzgibbon, Sean P. ;
Hughes, Emer ;
Rueckert, Daniel ;
Victor, Suresh ;
Rutherford, Mary ;
Edwards, A. David ;
Smith, Stephen M. ;
Tournier, Jacques-Donald ;
Hajnal, Joseph V. ;
Jbabdi, Saad ;
Sotiropoulos, Stamatios N. .
NEUROIMAGE, 2019, 185 :750-763
[7]   Deep Learning Techniques for Automatic MRI Cardiac Multi-Structures Segmentation and Diagnosis: Is the Problem Solved? [J].
Bernard, Olivier ;
Lalande, Alain ;
Zotti, Clement ;
Cervenansky, Frederick ;
Yang, Xin ;
Heng, Pheng-Ann ;
Cetin, Irem ;
Lekadir, Karim ;
Camara, Oscar ;
Gonzalez Ballester, Miguel Angel ;
Sanroma, Gerard ;
Napel, Sandy ;
Petersen, Steffen ;
Tziritas, Georgios ;
Grinias, Elias ;
Khened, Mahendra ;
Kollerathu, Varghese Alex ;
Krishnamurthi, Ganapathy ;
Rohe, Marc-Michel ;
Pennec, Xavier ;
Sermesant, Maxime ;
Isensee, Fabian ;
Jaeger, Paul ;
Maier-Hein, Klaus H. ;
Full, Peter M. ;
Wolf, Ivo ;
Engelhardt, Sandy ;
Baumgartner, Christian F. ;
Koch, Lisa M. ;
Wolterink, Jelmer M. ;
Isgum, Ivana ;
Jang, Yeonggul ;
Hong, Yoonmi ;
Patravali, Jay ;
Jain, Shubham ;
Humbert, Olivier ;
Jodoin, Pierre-Marc .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2018, 37 (11) :2514-2525
[8]  
Chen J., 2021, arXiv preprint arXiv:210204306, P04306
[9]  
Dehghani M., 2021, 9 INT C LEARN REPR I, DOI DOI 10.48550/ARXIV.2010.11929
[10]  
Dou H., 2020, ARXIV PREPRINT ARXIV