Exploring the user guidance for more accurate building segmentation from high-resolution remote sensing images

被引:2
作者
Yang, Dinghao [2 ]
Wang, Bin [2 ]
Li, Weijia [1 ]
He, Conghui [2 ]
机构
[1] Sun Yat Sen Univ, Sch Geospatial Engn & Sci, Guangzhou, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
User guidance; Building extraction; Semantic segmentation; Boundary correction; EXTRACTION;
D O I
10.1016/j.jag.2023.103609
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
In recent years, the computer vision domain has witnessed a surge of interest in interactive object segmentation, an area of study that seeks to expedite the annotation process for pixel-wise segmentation tasks through user guidance. Despite this growing focus, existing methods mainly focus on a single type of pre-annotation and neglect the quality of boundary prediction, which significantly influences subsequent manual adjustments to segmentation boundaries. To address these limitations, we introduce a novel end-to-end network to facilitate more precise building segmentation using diverse types of user guidance. In our proposed method, a centroid map is generated to provide foreground prior information crucial to the subsequent segmentation procedure, and the boundary correction module automatically refines the segmentation mask from existing segmentation networks. Extensive experiments on two popular building extraction datasets demonstrate that our method outperforms all existing approaches given various user guidance (bounding boxes, inside-outside points, or extreme points), achieving the IoU scores of over 95% on SpaceNet-Vegas dataset and over 93% on Inria-building dataset. The remarkable performance of our method further demonstrates its immense potential to alleviate the labor-intensive annotation process associated with remote sensing datasets. The code of our proposed method is available at https://github.com/StephenDHYang/UGBS-pytorch.
引用
收藏
页数:11
相关论文
共 48 条
[41]   A co-learning method to utilize optical images and photogrammetric point clouds for building extraction [J].
Xie, Yuxing ;
Tian, Jiaojiao ;
Zhu, Xiao Xiang .
INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2023, 116
[42]   CVNet: Contour Vibration Network for Building Extraction [J].
Xu, Ziqiang ;
Xu, Chunyan ;
Cui, Zhen ;
Zheng, Xiangwei ;
Yang, Jian .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :1373-1381
[43]  
Yang DH, 2023, Arxiv, DOI arXiv:2205.08324
[44]   DRE-Net: A Dynamic Radius-Encoding Neural Network with an Incremental Training Strategy for Interactive Segmentation of Remote Sensing Images [J].
Yang, Liangzhe ;
Zi, Wenjie ;
Chen, Hao ;
Peng, Shuang .
REMOTE SENSING, 2023, 15 (03)
[45]   SNNFD, spiking neural segmentation network in frequency domain using high spatial resolution images for building extraction [J].
Yu, Bo ;
Yang, Aqiang ;
Chen, Fang ;
Wang, Ning ;
Wang, Lei .
INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2022, 112
[46]  
Zhang SY, 2020, PROC CVPR IEEE, P12231, DOI [10.1109/TPAMI.2022.3227116, 10.1109/CVPR42600.2020.01225]
[47]   Pyramid Scene Parsing Network [J].
Zhao, Hengshuang ;
Shi, Jianping ;
Qi, Xiaojuan ;
Wang, Xiaogang ;
Jia, Jiaya .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6230-6239
[48]   Scene Parsing through ADE20K Dataset [J].
Zhou, Bolei ;
Zhao, Hang ;
Puig, Xavier ;
Fidler, Sanja ;
Barriuso, Adela ;
Torralba, Antonio .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5122-5130