Superpixel-Based Attention Graph Neural Network for Semantic Segmentation in Aerial Images

被引:21
作者
Diao, Qi [1 ]
Dai, Yaping [1 ]
Zhang, Ce [2 ,3 ]
Wu, Yan [4 ]
Feng, Xiaoxue [1 ]
Pan, Feng [1 ,5 ]
机构
[1] Beijing Inst Technol, Beijing 100081, Peoples R China
[2] Univ Lancaster, Lancaster Environm Ctr, Lancaster LA1 4YQ, England
[3] UK Ctr Ecol & Hydrol, Lib Ave, Lancaster LA1 4AP, England
[4] A STAR Inst Infocomm Res, Robot & Autonomous Syst Dept, Singapore 138632, Singapore
[5] Kunming BIT Ind Technol Res Inst Inc, Kunming 650106, Yunnan, Peoples R China
基金
中国国家自然科学基金;
关键词
graph neural networks; superpixel; attention mechanism; semantic segmentation; aerial images; FULLY CONVOLUTIONAL NETWORK; CLASSIFICATION; EXTRACTION;
D O I
10.3390/rs14020305
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Semantic segmentation is one of the significant tasks in understanding aerial images with high spatial resolution. Recently, Graph Neural Network (GNN) and attention mechanism have achieved excellent performance in semantic segmentation tasks in general images and been applied to aerial images. In this paper, we propose a novel Superpixel-based Attention Graph Neural Network (SAGNN) for semantic segmentation of high spatial resolution aerial images. A K-Nearest Neighbor (KNN) graph is constructed from our network for each image, where each node corresponds to a superpixel in the image and is associated with a hidden representation vector. On this basis, the initialization of the hidden representation vector is the appearance feature extracted by a unary Convolutional Neural Network (CNN) from the image. Moreover, relying on the attention mechanism and recursive functions, each node can update its hidden representation according to the current state and the incoming information from its neighbors. The final representation of each node is used to predict the semantic class of each superpixel. The attention mechanism enables graph nodes to differentially aggregate neighbor information, which can extract higher-quality features. Furthermore, the superpixels not only save computational resources, but also maintain object boundary to achieve more accurate predictions. The accuracy of our model on the Potsdam and Vaihingen public datasets exceeds all benchmark approaches, reaching 90.23% and 89.32%, respectively.
引用
收藏
页数:17
相关论文
共 57 条
  • [1] SLIC Superpixels Compared to State-of-the-Art Superpixel Methods
    Achanta, Radhakrishna
    Shaji, Appu
    Smith, Kevin
    Lucchi, Aurelien
    Fua, Pascal
    Suesstrunk, Sabine
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (11) : 2274 - 2281
  • [2] [Anonymous], 2014, INT C LEARN REPR
  • [3] Integrating Convolutional Neural Network and Multiresolution Segmentation for Land Cover and Land Use Mapping Using Satellite Imagery
    Atik, Saziye Ozge
    Ipbuker, Cengizhan
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (12):
  • [4] Superpixel Image Classification with Graph Attention Networks
    Avelar, Pedro H. C.
    Tavares, Anderson R.
    da Silveira, Thiago L. T.
    Jung, Cliudio R.
    Lamb, Luis C.
    [J]. 2020 33RD SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI 2020), 2020, : 203 - 209
  • [5] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [6] Bojchevski A, 2018, PR MACH LEARN RES, V80
  • [7] COCO-Stuff: Thing and Stuff Classes in Context
    Caesar, Holger
    Uijlings, Jasper
    Ferrari, Vittorio
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1209 - 1218
  • [8] DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs
    Chen, Liang-Chieh
    Papandreou, George
    Kokkinos, Iasonas
    Murphy, Kevin
    Yuille, Alan L.
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) : 834 - 848
  • [9] Chen LB, 2017, IEEE INT SYMP NANO, P1, DOI 10.1109/NANOARCH.2017.8053709
  • [10] The Cityscapes Dataset for Semantic Urban Scene Understanding
    Cordts, Marius
    Omran, Mohamed
    Ramos, Sebastian
    Rehfeld, Timo
    Enzweiler, Markus
    Benenson, Rodrigo
    Franke, Uwe
    Roth, Stefan
    Schiele, Bernt
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 3213 - 3223