MANet: A two-stage deep learning method for classification of COVID-19 from Chest X-ray images

被引:52
作者
Xu, Yujia [1 ]
Lam, Hak-Keung [1 ]
Jia, Guangyu [1 ]
机构
[1] Kings Coll London, Dept Engn, Ctr Robot Res, London WC2R 2LS, England
基金
英国工程与自然科学研究理事会;
关键词
Chest X-ray images; COVID-19; Two-stage; Segmentation; Spatial attention; Convolutional neural networks;
D O I
10.1016/j.neucom.2021.03.034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The early detection of infection is significant for the fight against the ongoing COVID-19 pandemic. Chest X-ray (CXR) imaging is an efficient screening technique via which lung infections can be detected. This paper aims to distinguish COVID-19 positive cases from the other four classes, including normal, tuberculosis (TB), bacterial pneumonia (BP), and viral pneumonia (VP), using CXR images. The existing COVID19 classification researches have achieved some successes with deep learning techniques while sometimes lacking interpretability and generalization ability. Hence, we propose a two-stage classification method MANet to address these issues in computer-aided COVID-19 diagnosis. Particularly, a segmentation model predicts the masks for all CXR images to extract their lung regions at the first stage. A followed classification CNN at the second stage then classifies the segmented CXR images into five classes based only on the preserved lung regions. In this segment-based classification task, we propose the mask attention mechanism (MA) which uses the predicted masks at the first stage as spatial attention maps to adjust the features of the CNN at the second stage. The MA spatial attention maps for features calculate the percentage of masked pixels in their receptive fields, suppressing the feature values based on the overlapping rates between their receptive fields and the segmented lung regions. In evaluation, we segment out the lung regions of all CXR images through a UNet with ResNet backbone, and then perform classification on the segmented CXR images using four classic CNNs with or without MA, including ResNet34, ResNet50, VGG16, and Inceptionv3. The experimental results illustrate that the classification models with MA have higher classification accuracy, more stable training process, and better interpretability and generalization ability than those without MA. Among the evaluated classification models, ResNet50 with MA achieves the highest average test accuracy of 96.32% in three runs, and the highest one is 97.06%. Meanwhile, the attention heat maps visualized by Grad-CAM indicate that models with MA make more reliable predictions based on the pathological patterns in lung regions. This further presents the potential of MANet to provide clinicians with diagnosis assistance. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:96 / 105
页数:10
相关论文
共 56 条
  • [1] Abbas A., 2021, APPL INTELL, DOI DOI 10.1007/S10489-020-01829-7
  • [2] Ai T, 2020, RADIOLOGY, DOI [10.5772/intechopen.80730, DOI 10.1148/radiol.2020200642, 10.1148/radiol.202020064224, DOI 10.1148/RADIOL.2020200642, 10.1148/radiol.2020200642]
  • [3] Covid-19: automatic detection from X-ray images utilizing transfer learning with convolutional neural networks
    Apostolopoulos, Ioannis D.
    Mpesiana, Tzani A.
    [J]. PHYSICAL AND ENGINEERING SCIENCES IN MEDICINE, 2020, 43 (02) : 635 - 640
  • [4] On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation
    Bach, Sebastian
    Binder, Alexander
    Montavon, Gregoire
    Klauschen, Frederick
    Mueller, Klaus-Robert
    Samek, Wojciech
    [J]. PLOS ONE, 2015, 10 (07):
  • [5] Bowles C., 2018, CoRR
  • [6] Albumentations: Fast and Flexible Image Augmentations
    Buslaev, Alexander
    Iglovikov, Vladimir I.
    Khvedchenya, Eugene
    Parinov, Alex
    Druzhinin, Mikhail
    Kalinin, Alexandr A.
    [J]. INFORMATION, 2020, 11 (02)
  • [7] Grad-CAM plus plus : Generalized Gradient-based Visual Explanations for Deep Convolutional Networks
    Chattopadhay, Aditya
    Sarkar, Anirban
    Howlader, Prantik
    Balasubramanian, Vineeth N.
    [J]. 2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, : 839 - 847
  • [8] DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs
    Chen, Liang-Chieh
    Papandreou, George
    Kokkinos, Iasonas
    Murphy, Kevin
    Yuille, Alan L.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 834 - 848
  • [9] Chen T, 2020, PR MACH LEARN RES, V119
  • [10] Exploring Simple Siamese Representation Learning
    Chen, Xinlei
    He, Kaiming
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15745 - 15753