A Deep Learning Network for Individual Tree Segmentation in UAV Images with a Coupled CSPNet and Attention Mechanism

被引:10
作者
Lv, Lujin [1 ,2 ,3 ]
Li, Xuejian [1 ,2 ,3 ]
Mao, Fangjie [1 ,2 ,3 ]
Zhou, Lv [4 ]
Xuan, Jie [1 ,2 ,3 ]
Zhao, Yinyin [1 ,2 ,3 ]
Yu, Jiacong [1 ,2 ,3 ]
Song, Meixuan [1 ,2 ,3 ]
Huang, Lei [1 ,2 ,3 ]
Du, Huaqiang [1 ,2 ,3 ]
机构
[1] Zhejiang A&F Univ, State Key Lab Subtrop Silviculture, Hangzhou 311300, Peoples R China
[2] Zhejiang A&F Univ, Key Lab Carbon Cycling Forest Ecosyst & Carbon Seq, Hangzhou 311300, Peoples R China
[3] Zhejiang A&F Univ, Sch Environm & Resources Sci, Hangzhou 311300, Peoples R China
[4] Beijing Forestry Univ, Res Ctr Forest Management Engn State Forestry & Gr, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
individual tree detection; Mask R-CNN; urban forest; deep learning; UAV; attention mechanism; CONVOLUTIONAL NEURAL-NETWORKS; CROWN DELINEATION; EXTRACTION; URBAN; CNN;
D O I
10.3390/rs15184420
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Accurate individual tree detection by unmanned aerial vehicles (UAVs) is a critical technique for smart forest management and serves as the foundation for evaluating ecological functions. Existing object detection and segmentation methods, on the other hand, have reduced accuracy when detecting and segmenting individual trees in complicated urban forest landscapes, as well as poor mask segmentation quality. This study proposes a novel Mask-CSP-attention-coupled network (MCAN) based on the Mask R-CNN algorithm. MCAN uses the Cross Stage Partial Net (CSPNet) framework with the Sigmoid Linear Unit (SiLU) activation function in the backbone network to form a new Cross Stage Partial Residual Net (CSPResNet) and employs a convolutional block attention module (CBAM) mechanism to the feature pyramid network (FPN) for feature fusion and multiscale segmentation to further improve the feature extraction ability of the model, enhance its detail information detection ability, and improve its individual tree detection accuracy. In this study, aerial photography of the study area was conducted by UAVs, and the acquired images were used to produce a dataset for training and validation. The method was compared with the Mask Region-based Convolutional Neural Network (Mask R-CNN), Faster Region-based Convolutional Neural Network (Faster R-CNN), and You Only Look Once v5 (YOLOv5) on the test set. In addition, four scenes-namely, a dense forest distribution, building forest intersection, street trees, and active plaza vegetation-were set up, and the improved segmentation network was used to perform individual tree segmentation on these scenes to test the large-scale segmentation ability of the model. MCAN's average precision (AP) value for individual tree identification is 92.40%, which is 3.7%, 3.84%, and 12.53% better than that of Mask R-CNN, Faster R-CNN, and YOLOv5, respectively. In comparison to Mask R-CNN, the segmentation AP value is 97.70%, an increase of 8.9%. The segmentation network's precision for the four scenes in multi-scene segmentation ranges from 95.55% to 92.33%, showing that the proposed network performs high-precision segmentation in many contexts.
引用
收藏
页数:19
相关论文
共 69 条
  • [1] Young and mature oil palm tree detection and counting using convolutional neural network deep learning method
    Abd Mubin, Nurulain
    Nadarajoo, Eiswary
    Shafri, Helmi Zulhaidi Mohd
    Hamedianfar, Alireza
    [J]. INTERNATIONAL JOURNAL OF REMOTE SENSING, 2019, 40 (19) : 7500 - 7515
  • [2] UAV-Based High Throughput Phenotyping in Citrus Utilizing Multispectral Imaging and Artificial Intelligence
    Ampatzidis, Yiannis
    Partel, Victor
    [J]. REMOTE SENSING, 2019, 11 (04)
  • [3] Quantifying tree cover in the forest-grassland ecotone of British Columbia using crown delineation and pattern detection
    Bai, Y
    Walsworth, N
    Roddan, B
    Hill, DA
    Broersma, K
    Thompson, D
    [J]. FOREST ECOLOGY AND MANAGEMENT, 2005, 212 (1-3) : 92 - 100
  • [4] Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
  • [5] An Enhanced Region Proposal Network for object detection using deep learning method
    Chen, Yu Peng
    Li, Ying
    Wang, Gang
    [J]. PLOS ONE, 2018, 13 (09):
  • [6] Mapping tree density at a global scale
    Crowther, T. W.
    Glick, H. B.
    Covey, K. R.
    Bettigole, C.
    Maynard, D. S.
    Thomas, S. M.
    Smith, J. R.
    Hintler, G.
    Duguid, M. C.
    Amatulli, G.
    Tuanmu, M. -N.
    Jetz, W.
    Salas, C.
    Stam, C.
    Piotto, D.
    Tavani, R.
    Green, S.
    Bruce, G.
    Williams, S. J.
    Wiser, S. K.
    Huber, M. O.
    Hengeveld, G. M.
    Nabuurs, G. -J.
    Tikhonova, E.
    Borchardt, P.
    Li, C. -F.
    Powrie, L. W.
    Fischer, M.
    Hemp, A.
    Homeier, J.
    Cho, P.
    Vibrans, A. C.
    Umunay, P. M.
    Piao, S. L.
    Rowe, C. W.
    Ashton, M. S.
    Crane, P. R.
    Bradford, M. A.
    [J]. NATURE, 2015, 525 (7568) : 201 - +
  • [7] Early Detection of Invasive Exotic Trees Using UAV and Manned Aircraft Multispectral and LiDAR Data
    Dash, Jonathan P.
    Watt, Michael S.
    Paul, Thomas S. H.
    Morgenroth, Justin
    Pearse, Grant D.
    [J]. REMOTE SENSING, 2019, 11 (15)
  • [8] Sigmoid-weighted linear units for neural network function approximation in reinforcement learning
    Elfwing, Stefan
    Uchibe, Eiji
    Doya, Kenji
    [J]. NEURAL NETWORKS, 2018, 107 : 3 - 11
  • [9] Swin-Transformer-Enabled YOLOv5 with Attention Mechanism for Small Object Detection on Satellite Images
    Gong, Hang
    Mu, Tingkui
    Li, Qiuxia
    Dai, Haishan
    Li, Chunlai
    He, Zhiping
    Wang, Wenjing
    Han, Feng
    Tuniyazi, Abudusalamu
    Li, Haoyang
    Lang, Xuechan
    Li, Zhiyuan
    Wang, Bin
    [J]. REMOTE SENSING, 2022, 14 (12)
  • [10] Guanhao Yang, 2020, 2020 IEEE 6th International Conference on Computer and Communications (ICCC), P1398, DOI 10.1109/ICCC51575.2020.9345042