Vision Transformers for Lung Segmentation on CXR Images

被引:7
作者
Ghali R. [1 ]
Akhloufi M.A. [1 ]
机构
[1] Perception, Robotics, and Intelligent Machines (PRIME), Department of Computer Science, Université de Moncton, Moncton, E1A 3E9, NB
基金
加拿大自然科学与工程研究理事会;
关键词
Chest X-rays; Deep learning; Lung segmentation; Medical image analysis; Vision transformers;
D O I
10.1007/s42979-023-01848-4
中图分类号
学科分类号
摘要
Accurate segmentation of the lungs in CXR images is the basis for an automated CXR image analysis system. It helps radiologists in detecting lung areas, subtle signs of disease and improving the diagnosis process for patients. However, precise semantic segmentation of lungs is considered a challenging case due to the presence of the edge rib cage, wide variation of lung shape, and lungs affected by diseases. In this paper, we address the problem of lung segmentation in healthy and unhealthy CXR images. Five models were developed and used in detecting and segmenting lung regions. Two loss functions and three benchmark datasets were employed to evaluate these models. Experimental results showed that the proposed models were able to extract salient global and local features from the input CXR images. The best performing model achieved an F1 score of 97.47%, outperforming recent published models. They proved their ability to separate lung regions from the rib cage and clavicle edges and segment varying lung shape depending on age and gender, as well as challenging cases of lungs affected by anomalies such as tuberculosis and the presence of nodules. © 2023, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
引用
收藏
相关论文
共 47 条
[41]  
Cohen J.P., Morrison P., Dao L., Roth K., Duong T.Q., Ghassemi M., Covid-19 Image Data Collection: Prospective Predictions are the Future, (2020)
[42]  
Singh A., Lall B., Panigrahi B.K., Agrawal A., Agrawal A., Thangakunam B., Christopher D.J., Deep lf-net: semantic lung segmentation from Indian chest radiographs including severely unhealthy images, Biomed Signal Process Control, 68, (2021)
[43]  
Chen L.-C., Zhu Y., Papandreou G., Schroff F., Adam H., Encoder-decoder with atrous separable convolution for semantic image segmentation, Computer Vision–ECCV 2018, pp. 833-851, (2018)
[44]  
Sandler M., Howard A., Zhu M., Zhmoginov A., Chen L.-C., Mobilenetv2: Inverted residuals and linear bottlenecks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4510-4520, (2018)
[45]  
Radosavovic I., Kosaraju R.P., Girshick R., He K., Dollar P., Designing network design spaces, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2020, pp. 10428-10436
[46]  
Paszke A., Gross S., Massa F., Lerer A., Bradbury J., Chanan G., Killeen T., Lin Z., Gimelshein N., Antiga L., Desmaison A., Kopf A., Yang E., Devito Z., Raison M., Tejani A., Chilamkurthy S., Steiner B., Fang L., Bai J., Chintala S.
[47]  
Deng J., Dong W., Socher R., Li L.-J., Li K., Fei-Fei L., Imagenet: A large-scale hierarchical image database, IEEE Conference on Computer Vision and Pattern Recognition, pp. 248-255, (2009)