UAV Imagery Real-Time Semantic Segmentation with Global-Local Information Attention

被引:0
作者
Zhang, Zikang [1 ]
Li, Gongquan [1 ]
机构
[1] Yangtze Univ, Sch Geosci, Wuhan 430100, Peoples R China
关键词
real-time semantic segmentation; drone imagery; feature fusion; global context information; NETWORK;
D O I
10.3390/s25061786
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
In real-time semantic segmentation for drone imagery, current lightweight algorithms suffer from the lack of integration of global and local information in the image, leading to missed detections and misclassifications in the classification categories. This paper proposes a method for the real-time semantic segmentation of drones that integrates multi-scale global context information. The principle utilizes a UNet structure, with the encoder employing a Resnet18 network to extract features. The decoder incorporates a global-local attention module, where the global branch compresses and extracts global information in both vertical and horizontal directions, and the local branch extracts local information through convolution, thereby enhancing the fusion of global and local information in the image. In the segmentation head, a shallow-feature fusion module is used to multi-scale integrate the various features extracted by the encoder, thereby strengthening the spatial information in the shallow features. The model was tested on the UAvid and UDD6 datasets, achieving accuracies of 68% mIoU (mean Intersection over Union) and 67% mIoU on the two datasets, respectively, 10% and 21.2% higher than the baseline model UNet. The real-time performance of the model reached 72.4 frames/s, which is 54.4 frames/s higher than the baseline model UNet. The experimental results demonstrate that the proposed model balances accuracy and real-time performance well.
引用
收藏
页数:19
相关论文
共 27 条
  • [1] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [2] A review on early wildfire detection from unmanned aerial vehicles using deep learning-based computer vision algorithms
    Bouguettaya, Abdelmalek
    Zarzour, Hafed
    Taberkit, Amine Mohammed
    Kechida, Ahmed
    [J]. SIGNAL PROCESSING, 2022, 190
  • [3] Chaurasia A, 2017, 2017 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP)
  • [4] Large-Scale Structure from Motion with Semantic Constraints of Aerial Images
    Chen, Yu
    Wang, Yao
    Lu, Peng
    Chen, Yisong
    Wang, Guoping
    [J]. PATTERN RECOGNITION AND COMPUTER VISION (PRCV 2018), PT I, 2018, 11256 : 347 - 359
  • [5] Comprehensive Semantic Segmentation on High Resolution UAV Imagery for Natural Disaster Damage Assessment
    Chowdhury, Tashnim
    Rahnemoonfar, Maryam
    Murphy, Robin
    Fernandes, Odair
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 3904 - 3913
  • [6] Rethinking BiSeNet For Real-time Semantic Segmentation
    Fan, Mingyuan
    Lai, Shenqi
    Huang, Junshi
    Wei, Xiaoming
    Chai, Zhenhua
    Luo, Junfeng
    Wei, Xiaolin
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9711 - 9720
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] Coordinate Attention for Efficient Mobile Network Design
    Hou, Qibin
    Zhou, Daquan
    Feng, Jiashi
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 13708 - 13717
  • [9] Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
  • [10] Real-Time Traffic Flow Parameter Estimation From UAV Video Based on Ensemble Classifier and Optical Flow
    Ke, Ruimin
    Li, Zhibin
    Tang, Jinjun
    Pan, Zewen
    Wang, Yinhai
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (01) : 54 - 64