A deep learning approach to DTM extraction from imagery using rule-based training labels

被引:51
作者
Gevaert, C. M. [1 ]
Persello, C. [1 ]
Nex, F. [1 ]
Vosselman, G. [1 ]
机构
[1] Univ Twente, ITC, Dept Earth Observat Sci, Enschede, Netherlands
关键词
Digital Terrain Models (DTM); Unmanned Aerial Vehicles (UAV); Aerial photogrammetry; Deep learning; Fully Convolutional Networks (FCN); BARE-EARTH EXTRACTION; BUILDING EXTRACTION; NEURAL-NETWORKS; AIRBORNE; CLASSIFICATION; SEGMENTATION; GENERATION; ACCURACY; FEATURES; MODELS;
D O I
10.1016/j.isprsjprs.2018.06.001
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
Existing algorithms for Digital Terrain Model (DTM) extraction still face difficulties due to data outliers and geometric ambiguities in the scene such as contiguous off-ground areas or sloped environments. We postulate that in such challenging cases, the radiometric information contained in aerial imagery may be leveraged to distinguish between ground and off-ground objects. We propose a method for DTM extraction from imagery which first applies morphological filters to the Digital Surface Model to obtain candidate ground and off-ground training samples. These samples are used to train a Fully Convolutional Network (FCN) in the second step, which can then be used to identify ground samples for the entire dataset. The proposed method harnesses the power of state-of-the-art deep learning methods, while showing how they can be adapted to the application of DTM extraction by (i) automatically selecting and labelling dataset-specific samples which can be used to train the network, and (ii) adapting the network architecture to consider a larger surface area without unnecessarily increasing the computational burden. The method is successfully tested on four datasets, indicating that the automatic labelling strategy can achieve an accuracy which is comparable to the use of manually labelled training samples. Furthermore, we demonstrate that the proposed method outperforms two reference DTM extraction algorithms in challenging areas.
引用
收藏
页码:106 / 123
页数:18
相关论文
共 60 条
[1]  
[Anonymous], 2018, IM2HEIGHT HEIGHT EST
[2]  
[Anonymous], OCTNETFUSION LEARNIN
[3]  
[Anonymous], ARXIV170306452
[4]   Semantic Segmentation of Earth Observation Data Using Multimodal and Multi-scale Deep Networks [J].
Audebert, Nicolas ;
Le Saux, Bertrand ;
Lefevre, Sebastien .
COMPUTER VISION - ACCV 2016, PT I, 2017, 10111 :180-196
[5]  
Axelsson P., 2000, INT ARCH PHOTOGRAMM, V33, P110, DOI DOI 10.1016/J.ISPRSJPRS.2005.10.005
[6]  
Bansal A., 2017, Pixelnet: Representation of the pixels, by the pixels, and for the pixels
[7]   Dense Object Reconstruction with Semantic Priors [J].
Bao, Sid Yingze ;
Chandraker, Manmohan ;
Lin, Yuanqing ;
Savarese, Silvio .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :1264-1271
[8]   Robust Optimization for Deep Regression [J].
Belagiannis, Vasileios ;
Rupprecht, Christian ;
Carneiro, Gustavo ;
Navab, Nassir .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2830-2838
[9]  
Beumier C, 2016, INT J REMOTE SENS, P1, DOI [10.1080/01431161.2016.1182666, DOI 10.1080/01431161.2016.1182666.]
[10]   Large-Scale Semantic 3D Reconstruction: an Adaptive Multi-Resolution Model for Multi-Class Volumetric Labeling [J].
Blaha, Maros ;
Vogel, Christoph ;
Richard, Audrey ;
Wegner, Jan D. ;
Pock, Thomas ;
Schindler, Konrad .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3176-3184