Evaluation of CNNs for Land Cover Classification in High Resolution Airborne Images

被引:6
作者
Haeufel, Gisela [1 ]
Lucks, Lukas [1 ]
Pohl, Melanie [1 ]
Bulatov, Dimitri [1 ]
Schilling, Hendrik [1 ]
机构
[1] Fraunhofer Inst Optron, Syst Technol & Image Exploitat IOSB, Gutleuthausstr 1, D-76275 Ettlingen, Germany
来源
EARTH RESOURCES AND ENVIRONMENTAL REMOTE SENSING/GIS APPLICATIONS IX | 2018年 / 10790卷
关键词
CNN; InceptionResNetV2; DeepLabV3+; land cover; image classification; superpixel; image recognition; image segmentation; SEGMENTATION;
D O I
10.1117/12.2325604
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Semantic land cover classification of satellite images or airborne images is becoming increasingly important for applications like urban planning, road net analysis or environmental monitoring. Sensor orientations or varying illumination make classification challenging. Depending on image source and classification task, it is not always easy to name the most discriminative features for a successful performance. To avoid feature selection, we transfer aspects of a feature-based classification approach to Convolutional Neural Networks (CNNs) which internally generate specific features. As land covering classes, we focus on buildings, roads, low (grass) and high vegetation (trees). Two different approaches will be analyzed: The first approach, using InceptionResNetV2, stems from networks used for image recognition. The second approach constitutes a fully convolutional neural network (DeepLabV3+) and is typically used for semantic image segmentation. Before processing, the image needs to be subdivided into tiles. This is necessary to make the data processible for the CNN, as well as for computational reasons. The tiles working with InceptionResNetV2 are derived from a superpixel segmentation. The tiles for working with DeepLabV3+ are overlapping tiles of a certain size. The advantages of this CNN is that its architecture enables to up-sample the classification result automatically and to produce a pixelwise labeling of the image content. As evaluation data for both approaches, we used the ISPRS benchmark of the city Vaihingen, Germany, containing true orthophotos and ground truth labeled for classification.
引用
收藏
页数:11
相关论文
共 38 条
[1]  
[Anonymous], 2017, ABS170605587 CORR
[2]  
[Anonymous], 2018, ECCV
[3]  
[Anonymous], 2014, Computer Science
[4]  
[Anonymous], ISPRS ANN PHOTOGRAMM
[5]  
[Anonymous], 2016, ARXIV161101962
[6]  
[Anonymous], 2017, P IEEE C COMP VIS PA
[7]  
[Anonymous], 2015, PROC CVPR IEEE
[8]  
[Anonymous], 2015, arXiv
[9]  
[Anonymous], 2017, ISPRS J PHOTOGRAMMET
[10]  
[Anonymous], 31 INT C MACH LEARN