U-Net-Id, an Instance Segmentation Model for Building Extraction from Satellite Images-Case Study in the Joanopolis City, Brazil

被引:42
作者
Wagner, Fabien H. [1 ,2 ]
Dalagnol, Ricardo [2 ]
Tarabalka, Yuliya [3 ,4 ]
Segantine, Tassiana Y. F. [1 ]
Thome, Rogerio [1 ]
Hirye, Mayumi C. M. [5 ]
机构
[1] Fdn Sci Technol & Space Applicat FUNCATE, GeoProc Div, BR-12210131 Sao Jose Dos Campos, SP, Brazil
[2] Natl Inst Space Res INPE, Remote Sensing Div, BR-12227010 Sao Jose Dos Campos, SP, Brazil
[3] Luxcarta Technol, Parc Activite Argile,Lot 119b, F-06370 Mouans Sartoux, France
[4] Inria Sophia Antipolis, F-06902 Sophia Antipolis, France
[5] Univ Sao Paulo, Fac Architecture & Urbanism, Quapa Lab, BR-05508900 Sao Paulo, SP, Brazil
基金
巴西圣保罗研究基金会; 英国自然环境研究理事会;
关键词
instance segmentation; U-net; building detection; urban landscape;
D O I
10.3390/rs12101544
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Currently, there exists a growing demand for individual building mapping in regions of rapid urban growth in less-developed countries. Most existing methods can segment buildings but cannot discriminate adjacent buildings. Here, we present a new convolutional neural network architecture (CNN) called U-net-id that performs building instance segmentation. The proposed network is trained with WorldView-3 satellite RGB images (0.3 m) and three different labeled masks. The first is the building mask; the second is the border mask, which is the border of the building segment with 4 pixels added outside and 3 pixels inside; and the third is the inner segment mask, which is the segment of the building diminished by 2 pixels. The architecture consists of three parallel paths, one for each mask, all starting with a U-net model. To accurately capture the overlap between the masks, all activation layers of the U-nets are copied and concatenated on each path and sent to two additional convolutional layers before the output activation layers. The method was tested with a dataset of 7563 manually delineated individual buildings of the city of Joanopolis-SP, Brazil. On this dataset, the semantic segmentation showed an overall accuracy of 97.67% and an F1-Score of 0.937 and the building individual instance segmentation showed good performance with a mean intersection over union (IoU) of 0.582 (median IoU = 0.694).
引用
收藏
页数:14
相关论文
共 36 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
Allaire J., 2019, keras: R Interface to Keras. R Package Version
[3]  
Allaire J., 2016, KERAS R INTERFACE KE
[4]  
[Anonymous], 2013, Core. R: A Language and Environment for Statistical Computing
[5]  
[Anonymous], 2020, SENSORS BASEL
[6]  
[Anonymous], 2017, P IEEE INT C COMPUTE
[7]   Deep Watershed Transform for Instance Segmentation [J].
Bai, Min ;
Urtasun, Raquel .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2858-2866
[8]   Cloud Detection for High-Resolution Satellite Imagery Using Machine Learning and Multi-Feature Fusion [J].
Bai, Ting ;
Li, Deren ;
Sun, Kaimin ;
Chen, Yepei ;
Li, Wenzhuo .
REMOTE SENSING, 2016, 8 (09)
[9]   Tree Crown Delineation Algorithm Based on a Convolutional Neural Network [J].
Braga, Jose R. G. ;
Peripato, Vinicius ;
Dalagnol, Ricardo ;
Ferreira, Matheus P. ;
Tarabalka, Yuliya ;
Aragao, Luiz E. O. C. ;
de Campos Velho, Haroldo E. ;
Shiguemori, Elcio H. ;
Wagner, Fabien H. .
REMOTE SENSING, 2020, 12 (08)
[10]   Uncovering Ecological Patterns with Convolutional Neural Networks [J].
Brodrick, Philip G. ;
Davies, Andrew B. ;
Asner, Gregory P. .
TRENDS IN ECOLOGY & EVOLUTION, 2019, 34 (08) :734-745