Wildfire Segmentation Using Deep Vision Transformers

被引:65
作者
Ghali, Rafik [1 ,2 ]
Akhloufi, Moulay A. [1 ]
Jmal, Marwa [3 ]
Mseddi, Wided Souidene [2 ]
Attia, Rabah [2 ]
机构
[1] Univ Moncton, Percept Robot & Intelligent Machines Res Grp PRIM, Dept Comp Sci, 18 Antonine Maillet Ave, Moncton, NB E1A 3E9, Canada
[2] Univ Carthage, Ecole Polytech Tunisie, SERCOM Lab, La Marsa 77-1054, Carthage, Tunisia
[3] Telnet Holding, Telnet Innovat Labs, Parc Elghazela Technol Commun, Ariana 2088, Tunisia
基金
加拿大自然科学与工程研究理事会;
关键词
forest fire detection; fire segmentation; vision Transformer; TransUNet; MedT; wildfires; FIRE-DETECTION; BURNED AREA; IMAGE; ALGORITHM; SEQUENCES; SEVERITY; NETWORK; COLOR;
D O I
10.3390/rs13173527
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
In this paper, we address the problem of forest fires' early detection and segmentation in order to predict their spread and help with fire fighting. Techniques based on Convolutional Networks are the most used and have proven to be efficient at solving such a problem. However, they remain limited in modeling the long-range relationship between objects in the image, due to the intrinsic locality of convolution operators. In order to overcome this drawback, Transformers, designed for sequence-to-sequence prediction, have emerged as alternative architectures. They have recently been used to determine the global dependencies between input and output sequences using the self-attention mechanism. In this context, we present in this work the very first study, which explores the potential of vision Transformers in the context of forest fire segmentation. Two vision-based Transformers are used, TransUNet and MedT. Thus, we design two frameworks based on the former image Transformers adapted to our complex, non-structured environment, which we evaluate using varying backbones and we optimize for forest fires' segmentation. Extensive evaluations of both frameworks revealed a performance superior to current methods. The proposed approaches achieved a state-of-the-art performance with an F1-score of 97.7% for TransUNet architecture and 96.0% for MedT architecture. The analysis of the results showed that these models reduce fire pixels mis-classifications thanks to the extraction of both global and local features, which provide finer detection of the fire's shape.
引用
收藏
页数:24
相关论文
共 96 条
[1]   Deep Learning Approaches Applied to Remote Sensing Datasets for Road Extraction: A State-Of-The-Art Review [J].
Abdollahi, Abolfazl ;
Pradhan, Biswajeet ;
Shukla, Nagesh ;
Chakraborty, Subrata ;
Alamri, Abdullah .
REMOTE SENSING, 2020, 12 (09)
[2]   Unsupervised Segmentation of Fire and Smoke From Infra-Red Videos [J].
Ajith, Meenu ;
Martinez-Ramon, Manel .
IEEE ACCESS, 2019, 7 :182381-182394
[3]   Wildland fires detection and segmentation using deep learning [J].
Akhloufi, Moulay A. ;
Tokime, Roger Booto ;
Elassady, Hassan .
PATTERN RECOGNITION AND TRACKING XXIX, 2018, 10649
[4]  
[Anonymous], P 3 INT C LEARNING R
[5]  
[Anonymous], 2019 10 INT WORKSH A
[6]   Long short-term memory [J].
Hochreiter, S ;
Schmidhuber, J .
NEURAL COMPUTATION, 1997, 9 (08) :1735-1780
[7]  
Bakator Mihalj, 2018, Multimodal Technologies and Interaction, V2, DOI 10.3390/mti2030047
[8]   wUUNet: Advanced Fully Convolutional Neural Network for Multiclass Fire Segmentation [J].
Bochkov, Vladimir Sergeevich ;
Kataeva, Liliya Yurievna .
SYMMETRY-BASEL, 2021, 13 (01) :1-18
[9]  
Bragilevsky L, 2017, IEEE PAC RIM CONF CO
[10]  
Carion Nicolas, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12346), P213, DOI 10.1007/978-3-030-58452-8_13