Exploring the user guidance for more accurate building segmentation from high-resolution remote sensing images

被引:2
作者
Yang, Dinghao [2 ]
Wang, Bin [2 ]
Li, Weijia [1 ]
He, Conghui [2 ]
机构
[1] Sun Yat Sen Univ, Sch Geospatial Engn & Sci, Guangzhou, Peoples R China
[2] Shanghai Artificial Intelligence Lab, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
User guidance; Building extraction; Semantic segmentation; Boundary correction; EXTRACTION;
D O I
10.1016/j.jag.2023.103609
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
In recent years, the computer vision domain has witnessed a surge of interest in interactive object segmentation, an area of study that seeks to expedite the annotation process for pixel-wise segmentation tasks through user guidance. Despite this growing focus, existing methods mainly focus on a single type of pre-annotation and neglect the quality of boundary prediction, which significantly influences subsequent manual adjustments to segmentation boundaries. To address these limitations, we introduce a novel end-to-end network to facilitate more precise building segmentation using diverse types of user guidance. In our proposed method, a centroid map is generated to provide foreground prior information crucial to the subsequent segmentation procedure, and the boundary correction module automatically refines the segmentation mask from existing segmentation networks. Extensive experiments on two popular building extraction datasets demonstrate that our method outperforms all existing approaches given various user guidance (bounding boxes, inside-outside points, or extreme points), achieving the IoU scores of over 95% on SpaceNet-Vegas dataset and over 93% on Inria-building dataset. The remarkable performance of our method further demonstrates its immense potential to alleviate the labor-intensive annotation process associated with remote sensing datasets. The code of our proposed method is available at https://github.com/StephenDHYang/UGBS-pytorch.
引用
收藏
页数:11
相关论文
共 48 条
[1]   Efficient Interactive Annotation of Segmentation Datasets with Polygon-RNN plus [J].
Acuna, David ;
Ling, Huan ;
Kar, Amlan ;
Fidler, Sanja .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :859-868
[2]  
[Anonymous], 2017, BMVC
[3]   Annotating Object Instances with a Polygon-RNN [J].
Castrejon, Lluis ;
Kundu, Kaustav ;
Urtasun, Raquel ;
Fidler, Sanja .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4485-4493
[4]  
Chen L.-C., P EUROPEAN C COMPUT, P801
[5]   FocalClick: Towards Practical Interactive Image Segmentation [J].
Chen, Xi ;
Zhao, Zhiyan ;
Zhang, Yilei ;
Duan, Manni ;
Qi, Donglian ;
Zhao, Hengshuang .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :1290-1299
[6]   Cascaded Pyramid Network for Multi-Person Pose Estimation [J].
Chen, Yilun ;
Wang, Zhicheng ;
Peng, Yuxiang ;
Zhang, Zhiqiang ;
Yu, Gang ;
Sun, Jian .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7103-7112
[7]   Boundary IoU: Improving Object-Centric Image Segmentation Evaluation [J].
Cheng, Bowen ;
Girshick, Ross ;
Dollar, Piotr ;
Berg, Alexander C. ;
Kirillov, Alexander .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :15329-15337
[8]   DARNet: Deep Active Ray Network for Building Segmentation [J].
Cheng, Dominic ;
Liao, Renjie ;
Fidler, Sanja ;
Urtasun, Raquel .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :7423-7431
[9]   CascadePSP: Toward Class-Agnostic and Very High-Resolution Segmentation via Global and Local Refinement [J].
Cheng, Ho Kei ;
Chung, Jihoon ;
Tai, Yu-Wing ;
Tang, Chi-Keung .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :8887-8896
[10]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338