Mask R-CNN Refitting Strategy for Plant Counting and Sizing in UAV Imagery

被引:81
作者
Machefer, Melissande [1 ,2 ]
Lemarchand, Francois [1 ]
Bonnefond, Virginie [1 ]
Hitchins, Alasdair [1 ]
Sidiropoulos, Panagiotis [1 ,3 ]
机构
[1] Hummingbird Technol, Aviat House,125 Kingsway, London WC2B 6NH, England
[2] Lobelia IsardSAT, Technol Pk,8-14 Marie Curie St, Barcelona 08042, Spain
[3] UCL, Mullard Space Sci Lab, London WC1E 6BT, England
关键词
UAV; crop mapping; image analysis; precision agriculture; deep learning; individual plant segmentation; plant detection; transfer learning; SEGMENTATION;
D O I
10.3390/rs12183015
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
This work introduces a method that combines remote sensing and deep learning into a framework that is tailored for accurate, reliable and efficient counting and sizing of plants in aerial images. The investigated task focuses on two low-density crops, potato and lettuce. This double objective of counting and sizing is achieved through the detection and segmentation of individual plants by fine-tuning an existing deep learning architecture called Mask R-CNN. This paper includes a thorough discussion on the optimal parametrisation to adapt the Mask R-CNN architecture to this novel task. As we examine the correlation of the Mask R-CNN performance to the annotation volume and granularity (coarse or refined) of remotely sensed images of plants, we conclude that transfer learning can be effectively used to reduce the required amount of labelled data. Indeed, a previously trained Mask R-CNN on a low-density crop can improve performances after training on new crops. Once trained for a given crop, the Mask R-CNN solution is shown to outperform a manually-tuned computer vision algorithm. Model performances are assessed using intuitive metrics such as Mean Average Precision (mAP) from Intersection over Union (IoU) of the masks for individual plant segmentation and Multiple Object Tracking Accuracy (MOTA) for detection. The presented model reaches an mAP of0.418for potato plants and0.660for lettuces for the individual plant segmentation task. In detection, we obtain a MOTA of0.781for potato plants and0.918for lettuces.
引用
收藏
页数:23
相关论文
共 58 条
[41]  
Redmon J., 2016, PROC IEEE C COMPUTER, p7263 7271
[42]  
Redmon J, 2018, Arxiv, DOI [arXiv:1804.02767, DOI 10.48550/ARXIV.1804.02767]
[43]   You Only Look Once: Unified, Real-Time Object Detection [J].
Redmon, Joseph ;
Divvala, Santosh ;
Girshick, Ross ;
Farhadi, Ali .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :779-788
[44]  
Rees S., 2011, AUST J MULTIDISCIPLI, V8, P97, DOI [10.1080/14488388.2011.11464829, DOI 10.1080/14488388.2011.11464829]
[45]   Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks [J].
Ren, Shaoqing ;
He, Kaiming ;
Girshick, Ross ;
Sun, Jian .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) :1137-1149
[46]  
Ribera J, 2017, IEEE GLOB CONF SIG, P1344, DOI 10.1109/GlobalSIP.2017.8309180
[47]   U-Net: Convolutional Networks for Biomedical Image Segmentation [J].
Ronneberger, Olaf ;
Fischer, Philipp ;
Brox, Thomas .
MEDICAL IMAGE COMPUTING AND COMPUTER-ASSISTED INTERVENTION, PT III, 2015, 9351 :234-241
[48]  
Simonyan K, 2015, Arxiv, DOI arXiv:1409.1556
[49]   Kiwifruit detection in field images using Faster R-CNN with VGG16 [J].
Song, Zhenzhen ;
Fu, Longsheng ;
Wu, Jingzhu ;
Liu, Zhihao ;
Li, Rui ;
Cui, Yongjie .
IFAC PAPERSONLINE, 2019, 52 (30) :76-81
[50]  
Szegedy C., ARXIV160207261