Transformer Neural Network for Weed and Crop Classification of High Resolution UAV Images

被引:108
|
作者
Reedha, Reenul [1 ]
Dericquebourg, Eric [1 ]
Canals, Raphael [2 ]
Hafiane, Adel [1 ]
机构
[1] Univ Orleans, INSA CVL, PRISME Lab EA 4229, F-18022 Bourges, France
[2] Univ Orleans, PRISME Lab EA 4229, INSA CVL, F-45067 Orleans, France
关键词
computer vision; deep learning; self-attention; vision transformers; remote sensing; drone; image classification; agriculture;
D O I
10.3390/rs14030592
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Monitoring crops and weeds is a major challenge in agriculture and food production today. Weeds compete directly with crops for moisture, nutrients, and sunlight. They therefore have a significant negative impact on crop yield if not sufficiently controlled. Weed detection and mapping is an essential step in weed control. Many existing research studies recognize the importance of remote sensing systems and machine learning algorithms in weed management. Deep learning approaches have shown good performance in many agriculture-related remote sensing tasks, such as plant classification, disease detection, etc. However, despite the success of these approaches, they still face many challenges such as high computation cost, the need of large labelled datasets, intra-class discrimination (in growing phase weeds and crops share many attributes similarity as color, texture, and shape), etc. This paper aims to show that the attention-based deep network is a promising approach to address the forementioned problems, in the context of weeds and crops recognition with drone system. The specific objective of this study was to investigate visual transformers (ViT) and apply them to plant classification in Unmanned Aerial Vehicles (UAV) images. Data were collected using a high-resolution camera mounted on a UAV, which was deployed in beet, parsley and spinach fields. The acquired data were augmented to build larger dataset, since ViT requires large sample sets for better performance, we also adopted the transfer learning strategy. Experiments were set out to assess the effect of training and validation dataset size, as well as the effect of increasing the test set while reducing the training set. The results show that with a small labeled training dataset, the ViT models outperform state-of-the-art models such as EfficientNet and ResNet. The results of this study are promising and show the potential of ViT to be applied to a wide range of remote sensing image analysis tasks.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Transformer Faults Classification Based on Convolution Neural Network
    Elmohallawy, Maha A.
    Abdel-Gawad, Amal F.
    Hassan, Amir Yassin
    Selem, Sameh I.
    INTERNATIONAL JOURNAL OF ELECTRICAL AND COMPUTER ENGINEERING SYSTEMS, 2023, 14 (09) : 1069 - 1075
  • [32] Potential of Ultra-High-Resolution UAV Images with Centimeter GNSS Positioning for Plant Scale Crop Monitoring
    Gilliot, Jean-Marc
    Hadjar, Dalila
    Michelin, Joel
    REMOTE SENSING, 2022, 14 (10)
  • [33] Joint superpixel and Transformer for high resolution remote sensing image classification
    Guangpu Dang
    Zhongan Mao
    Tingyu Zhang
    Tao Liu
    Tao Wang
    Liangzhi Li
    Yu Gao
    Runqing Tian
    Kun Wang
    Ling Han
    Scientific Reports, 14
  • [34] Road Damage Detection Using the Hunger Games Search with Elman Neural Network on High-Resolution Remote Sensing Images
    Al Duhayyim, Mesfer
    Malibari, Areej A.
    Alharbi, Abdullah
    Afef, Kallekh
    Yafoz, Ayman
    Alsini, Raed
    Alghushairy, Omar
    Mohsen, Heba
    REMOTE SENSING, 2022, 14 (24)
  • [35] Joint superpixel and Transformer for high resolution remote sensing image classification
    Dang, Guangpu
    Mao, Zhongan
    Zhang, Tingyu
    Liu, Tao
    Wang, Tao
    Li, Liangzhi
    Gao, Yu
    Tian, Runqing
    Wang, Kun
    Han, Ling
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [36] Crop extraction combined with airborne hyperspectral images and deep convolutional neural network
    Li, Qiang
    Shao, Zhitao
    Liu, Yapin
    Li, Tian
    Wang, Xiyao
    Yu, Junchuan
    AOPC 2021: OPTICAL SENSING AND IMAGING TECHNOLOGY, 2021, 12065
  • [37] Hybrid CNN and Transformer Network for Semantic Segmentation of UAV Remote Sensing Images
    Zhou X.
    Zhou L.
    Gong S.
    Zhang H.
    Zhong S.
    Xia Y.
    Huang Y.
    IEEE Journal on Miniaturization for Air and Space Systems, 2024, 5 (01): : 33 - 41
  • [38] SCENE CLASSIFICATION OF HIGH RESOLUTION REMOTE SENSING IMAGES USING CONVOLUTIONAL NEURAL NETWORKS
    Cheng, Gong
    Ma, Chengcheng
    Zhou, Peicheng
    Yao, Xiwen
    Han, Junwei
    2016 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2016, : 767 - 770
  • [39] SwinT-SRNet: Swin transformer with image super-resolution reconstruction network for pollen images classification
    Zu, Baokai
    Cao, Tong
    Li, Yafang
    Li, Jianqiang
    Ju, Fujiao
    Wang, Hongyuan
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [40] Spatial-Spectral Neural Network for High Resolution Multispectral Image Classification
    Tanha, Roozbeh
    Ghassemian, Hassan
    PROCEEDINGS OF THE 13TH IRANIAN/3RD INTERNATIONAL MACHINE VISION AND IMAGE PROCESSING CONFERENCE, MVIP, 2024, : 51 - 55