Mask R-CNN Refitting Strategy for Plant Counting and Sizing in UAV Imagery

被引:81
作者
Machefer, Melissande [1 ,2 ]
Lemarchand, Francois [1 ]
Bonnefond, Virginie [1 ]
Hitchins, Alasdair [1 ]
Sidiropoulos, Panagiotis [1 ,3 ]
机构
[1] Hummingbird Technol, Aviat House,125 Kingsway, London WC2B 6NH, England
[2] Lobelia IsardSAT, Technol Pk,8-14 Marie Curie St, Barcelona 08042, Spain
[3] UCL, Mullard Space Sci Lab, London WC1E 6BT, England
关键词
UAV; crop mapping; image analysis; precision agriculture; deep learning; individual plant segmentation; plant detection; transfer learning; SEGMENTATION;
D O I
10.3390/rs12183015
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
This work introduces a method that combines remote sensing and deep learning into a framework that is tailored for accurate, reliable and efficient counting and sizing of plants in aerial images. The investigated task focuses on two low-density crops, potato and lettuce. This double objective of counting and sizing is achieved through the detection and segmentation of individual plants by fine-tuning an existing deep learning architecture called Mask R-CNN. This paper includes a thorough discussion on the optimal parametrisation to adapt the Mask R-CNN architecture to this novel task. As we examine the correlation of the Mask R-CNN performance to the annotation volume and granularity (coarse or refined) of remotely sensed images of plants, we conclude that transfer learning can be effectively used to reduce the required amount of labelled data. Indeed, a previously trained Mask R-CNN on a low-density crop can improve performances after training on new crops. Once trained for a given crop, the Mask R-CNN solution is shown to outperform a manually-tuned computer vision algorithm. Model performances are assessed using intuitive metrics such as Mean Average Precision (mAP) from Intersection over Union (IoU) of the masks for individual plant segmentation and Multiple Object Tracking Accuracy (MOTA) for detection. The presented model reaches an mAP of0.418for potato plants and0.660for lettuces for the individual plant segmentation task. In detection, we obtain a MOTA of0.781for potato plants and0.918for lettuces.
引用
收藏
页数:23
相关论文
共 58 条
[1]  
Abdulla W., 2017, Mask R-CNN for object detection and instance segmentation on Keras and TensorFlow
[2]   DeepWheat: Estimating Phenotypic Traits from Crop Images with Deep Learning [J].
Aich, Shubhra ;
Josuttes, Anique ;
Ovsyannikov, Ilya ;
Strueby, Keegan ;
Ahmed, Imran ;
Duddu, Hema Sudhakar ;
Pozniak, Curtis ;
Shirtliffe, Steve ;
Stavness, Ian .
2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, :323-332
[3]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[4]   Evaluating Multiple Object Tracking Performance: The CLEAR MOT Metrics [J].
Bernardin, Keni ;
Stiefelhagen, Rainer .
EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2008, 2008 (1)
[5]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[6]  
Chen X., ARXIV 2017 170202138
[7]   CentroidNet: A Deep Neural Network for Joint Object Localization and Counting [J].
Dijkstra, K. ;
van de Loosdrecht, J. ;
Schomaker, L. R. B. ;
Wiering, M. A. .
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2018, PT III, 2019, 11053 :585-601
[8]  
Epperson M., 2014, TECHNICAL REPORT
[9]   Automatic Tobacco Plant Detection in UAV Images via Deep Neural Networks [J].
Fan, Zhun ;
Lu, Jiewei ;
Gong, Maoguo ;
Xie, Honghui ;
Goodman, Erik D. .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2018, 11 (03) :876-887
[10]   Deep Orange: Mask R-CNN based Orange Detection and Segmentation [J].
Ganesh, P. ;
Volle, K. ;
Burks, T. F. ;
Mehta, S. S. .
IFAC PAPERSONLINE, 2019, 52 (30) :70-75