A Lightweight Deep Learning Architecture for Vegetation Segmentation using UAV-captured Aerial Images

被引:23
作者
Behera, Tanmay Kumar [1 ]
Bakshi, Sambit [1 ]
Sa, Pankaj Kumar [1 ]
机构
[1] Natl Inst Technol Rourkela, Dept Comp Sci & Engn, Rourkela 769008, Odisha, India
关键词
Remote sensing; Semantic segmentation; Depthwise separable convolution; Deep learning; Convolutional neural network (CNN); Urban mapping; Vegetation segmentation; Unmanned aerial vehicle (UAV); ENVIRONMENT; NETWORK;
D O I
10.1016/j.suscom.2022.100841
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The unmanned aerial vehicle (UAV)-captured panoptic remote sensing images have great potential to promote robotics-inspired intelligent solutions for land cover mapping, disaster management, smart agriculture through automatic vegetation detection, and real-time environmental surveillance. However, many of these applications require faster execution of the tasks to get the job done in real-time. In this regard, this article proposes a lightweight convolutional neural network (CNN) architecture, also known as LW-AerialSegNet, that helps preserve the network's feed-forward nature by increasing the intermediate layers to gather more crucial features for the segmentation task. Moreover, the network uses the concept of densely connected architecture and depth-wise separable convolution mechanisms to reduce the number of parameters of the model that can be deployed in the internet of things (IoT) edge devices to perform real-time segmentation. A UAV-based image segmentation dataset NITRDrone Dataset and Urban Drone dataset (UDD) are used to evaluate the proposed architecture. It has achieved an intersection over union (IoU) of 82% and 71% on the NITRDrone datasets UDD, respectively, thereby illustrating its superiority among the considered state-of-the-art mechanisms. The experimental results indicate that the implementation of depth-wise separable convolutions helps reduce the number of trainable parameters significantly, making it suitable to be applied on edge-computing devices at a smaller scale. The proposed architecture can be deployed in real-life settings on a UAV to extract objects such as vegetation and road lines, hence can be used in mapping urban areas, agricultural lands, etc.
引用
收藏
页数:11
相关论文
共 53 条
[1]  
Agarap A F., 2018, Deep learning using rectified linear units (relu)
[2]  
[Anonymous], 2016, PYTORCH DOCUMENTS
[3]  
[Anonymous], 2022, SATELLITE STAT
[4]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[5]   Vegetation Extraction from UAV-based Aerial Images through Deep Learning [J].
Behera, Tanmay Kumar ;
Bakshi, Sambit ;
Sa, Pankaj Kumar .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2022, 198
[6]   The NITRDrone Dataset to Address the Challenges for Road Extraction from Aerial Images [J].
Behera, Tanmay Kumar ;
Bakshi, Sambit ;
Sa, Pankaj Kumar ;
Nappi, Michele ;
Castiglione, Aniello ;
Vijayakumar, Pandi ;
Gupta, Brij Bhooshan .
JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2023, 95 (2-3) :197-209
[7]   Aerial Data Aiding Smart Societal Reformation: Current Applications and Path Ahead [J].
Behera, Tanmay Kumar ;
Bakshi, Sambit ;
Sa, Pankaj Kumar .
IT PROFESSIONAL, 2021, 23 (03) :82-87
[8]   Collection of Data With Drones in Precision Agriculture: Analytical Model and LoRa Case Study [J].
Caruso, Antonio ;
Chessa, Stefano ;
Escolar, Soledad ;
Barba, Jesus ;
Lopez, Juan Carlos .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (22) :16692-16704
[9]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[10]   Large-Scale Structure from Motion with Semantic Constraints of Aerial Images [J].
Chen, Yu ;
Wang, Yao ;
Lu, Peng ;
Chen, Yisong ;
Wang, Guoping .
PATTERN RECOGNITION AND COMPUTER VISION (PRCV 2018), PT I, 2018, 11256 :347-359