Deep learning on edge: Extracting field boundaries from satellite images with a convolutional neural network

被引:188
作者
Waldner, Francois [1 ]
Diakogiannis, Foivos, I [2 ]
机构
[1] CSIRO, Agr & Food, 306 Carmody Rd, St Lucia, Qld, Australia
[2] CSIRO, Data61, Analyt, 147 Underwood Ave, Floreat, WA, Australia
关键词
Agriculture; Field boundaries; Sentinel-2; Semantic segmentation; Instance segmentation; Multitasking; Computer vision; Generalisation; RESOLUTION; DELINEATION; SEGMENTATION; CLASSIFICATION; EFFICIENCY; ACCURACY; OBJECTS;
D O I
10.1016/j.rse.2020.111741
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Applications of digital agricultural services often require either farmers or their advisers to provide digital records of their field boundaries. Automatic extraction of field boundaries from satellite imagery would reduce the reliance on manual input of these records, which is time consuming, and would underpin the provision of remote products and services. The lack of current field boundary data sets seems to indicate low uptake of existing methods, presumably because of expensive image preprocessing requirements and local, often arbitrary, tuning. In this paper, we propose a data-driven, robust and general method to facilitate field boundary extraction from satellite images. We formulated this task as a multi-task semantic segmentation problem. We used ResUNet-a, a deep convolutional neural network with a fully connected UNet backbone that features dilated convolutions and conditioned inference to identify: 1) the extent of fields; 2) the field boundaries; and 3) the distance to the closest boundary. By asking the algorithm to reconstruct three correlated outputs, the model's performance and its ability to generalise greatly improve. Segmentation of individual fields was then achieved by post-processing the three model outputs, e.g., via thresholding or watershed segmentation. Using a single monthly composite image from Sentinel-2 as input, our model was highly accurate in mapping field extent, field boundaries and, consequently, individual fields. Replacing the monthly composite with a single-date image close to the compositing period marginally decreased accuracy. We then showed in a series of experiments that, without recalibration, the same model generalised well across resolutions (10 m to 30 m), sensors (Sentinel-2 to Landsat-8), space and time. Building consensus by averaging model predictions from at least four images acquired across the season is paramount to reducing the temporal variations of accuracy. Our convolutional neural network is capable of learning complex hierarchical contextual features from the image to accurately detect field boundaries and discard irrelevant boundaries, thereby outperforming conventional edge filters. By minimising over-fitting and image preprocessing requirements, and by replacing local arbitrary decisions by data-driven ones, our approach is expected to facilitate the extraction of individual crop fields at scale.
引用
收藏
页数:15
相关论文
共 63 条
[1]  
[Anonymous], FIELD CROP BOUND DAT
[2]  
[Anonymous], ABS170108816 CORR
[3]  
[Anonymous], 2018, Soft Computing Based Medical Image Analysis eds
[4]  
[Anonymous], 2017, NASA MAKING EARTH SY, DOI DOI 10.5067/MEASURES/GFSAD/GFSAD30NACE.001
[5]  
[Anonymous], 2017, NASA Making Earth System Data Records for Use in Research Environments (MEaSUREs) Global Food Security-support Analysis Data (GFSAD) Cropland Extent 2015 South America 30 m V001, DOI DOI 10.5067/MEASURES/GFSAD/GFSAD30SACE.001
[6]  
[Anonymous], 2012, International Journal of Computer Science Issues (IJCSI)
[7]   Sentinel-2 cropland mapping using pixel-based and object-based time-weighted dynamic time warping analysis [J].
Belgiu, Mariana ;
Csillik, Ovidiu .
REMOTE SENSING OF ENVIRONMENT, 2018, 204 :509-523
[8]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[9]   Efficiency of crop identification based on optical and SAR image time series [J].
Blaes, X ;
Vanhalle, L ;
Defourny, P .
REMOTE SENSING OF ENVIRONMENT, 2005, 96 (3-4) :352-365
[10]   DISTANCE TRANSFORMATIONS IN DIGITAL IMAGES [J].
BORGEFORS, G .
COMPUTER VISION GRAPHICS AND IMAGE PROCESSING, 1986, 34 (03) :344-371