Rse-net: Road-shape enhanced neural network for Road extraction in high resolution remote sensing image

被引:5
作者
Bai, Xiangtian [1 ]
Guo, Li [1 ]
Huo, Hongyuan [2 ]
Zhang, Jiangshui [1 ]
Zhang, Yi [1 ]
Li, Zhao-Liang [3 ,4 ]
机构
[1] Informat Engn Univ, Inst Geospatial Informat, Zhengzhou 450001, Peoples R China
[2] Beijing Univ Technol, Fac Architecture Civil & Transportat Engn, Beijing, Peoples R China
[3] Chinese Agr Acad Sci, Minist Agr, Inst Agr Resources & Reg Planning, Key Lab Agr Remote Sensing, Beijing, Peoples R China
[4] Univ Strasbourg, ICube, CNRS, Illkirch Graffenstaden, France
关键词
road extraction; semantic segmentation; deep learning; Convolutional Neural Network; multi-scale; AWARE;
D O I
10.1080/01431161.2023.2214277
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Automatically extracting roads from remote sensing images is increasingly important for autonomous driving and geographic information systems. However, due to factors such as ground objects blocking road boundaries and some roads being narrow and long in high-resolution remote sensing images, it is difficult for the current mainstream pixel-level extraction model to guarantee the continuity of roads and the smoothness of boundaries. To solve this problem, this paper designs a neural network Rse-net based on semantic segmentation, using a two-stream semantic segmentation network algorithm to make the network pay more attention to the boundary information of roads and narrow roads. This structure uses shape stream to process road boundary information (roads' shape stream), and processes it in parallel with classic stream. Use the Gated Shape CNN to connect the two streams of roads' shape stream and classic stream, and use the classical stream to optimize the boundary information in the shape stream. At the same time, a multi-scale convolutional attention mechanism is used between the decoder and the encoder to expand the receptive field through large-core attention, and obtain more semantic information without causing too much calculation. Finally, the effectiveness of the proposed network is verified by comparative experiments.
引用
收藏
页码:7339 / 7360
页数:22
相关论文
共 27 条
  • [1] Abdelfattah R., 2022, ARXIV PREPRINT ARXIV, V2204, P07243, DOI [https://doi.org/10.48550/arXiv.2204.07243, DOI 10.48550/ARXIV.2204.07243]
  • [2] RoadVecNet: a new approach for simultaneous road network segmentation and vectorization from aerial and google earth imagery in a complex urban set-up
    Abdollahi, Abolfazl
    Pradhan, Biswajeet
    Alamri, Abdullah
    [J]. GISCIENCE & REMOTE SENSING, 2021, 58 (07) : 1151 - 1174
  • [3] Deep Learning Approaches Applied to Remote Sensing Datasets for Road Extraction: A State-Of-The-Art Review
    Abdollahi, Abolfazl
    Pradhan, Biswajeet
    Shukla, Nagesh
    Chakraborty, Subrata
    Alamri, Abdullah
    [J]. REMOTE SENSING, 2020, 12 (09)
  • [4] Building Extraction from Remote Sensing Images with Sparse Token Transformers
    Chen, Keyan
    Zou, Zhengxia
    Shi, Zhenwei
    [J]. REMOTE SENSING, 2021, 13 (21)
  • [5] GA-Net: A geometry prior assisted neural network for road extraction
    Chen, Xin
    Sun, Qun
    Guo, Wenyue
    Qiu, Chunping
    Yu, Anzhu
    [J]. INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2022, 114
  • [6] Gao H., 2019, CHIN HIGH RES EARTH, P187
  • [7] Amelotin Gene Structure and Expression during Enamel Formation in the Opossum Monodelphis domestica
    Gasse, Barbara
    Liu, Xi
    Corre, Erwan
    Sire, Jean-Yves
    [J]. PLOS ONE, 2015, 10 (07):
  • [8] A road extraction methodforintelligent dispatchingbased on MD-LinkNeSt network in open-pit mine
    Gu, Qinghua
    Xue, Buqing
    Ruan, Shunling
    Li, Xuexian
    [J]. INTERNATIONAL JOURNAL OF MINING RECLAMATION AND ENVIRONMENT, 2021, 35 (09) : 656 - 669
  • [9] Guo M-H., 2022, ADV NEURAL INF PROCE, V35, P1140, DOI DOI 10.48550/ARXIV.2209.08575
  • [10] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778