Fire Detection Approach Based on Vision Transformer

被引:7
|
作者
Khudayberdiev, Otabek [1 ]
Zhang, Jiashu [1 ]
Elkhalil, Ahmed [1 ]
Balde, Lansana [1 ]
机构
[1] Southwest Jiaotong Univ, Sch Comp & Artificial Intelligence, Chengdu, Peoples R China
来源
ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT I | 2022年 / 13338卷
关键词
Vision transformer; Self-attention; Convolutional neural networks; Fire detection; Image classification; CONVOLUTIONAL NEURAL-NETWORKS; SURVEILLANCE;
D O I
10.1007/978-3-031-06794-5_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Considering the rapid development of embedding surveillance video systems for fire monitoring, we need to distribute systems with high accuracy and detection speed. Recent progress in vision-based fire detection techniques achieved remarkable success by the powerful ability of deep convolutional neural networks. CNN's have long been the architecture of choice for computer vision tasks. However, current CNN-based methods consider fire classification entire image pixels as equal, ignoring regardless of information. Thus, this can cause a low accuracy rate and delay detection. To increase detection speed and achieve high accuracy, we propose a fire detection approach based on Vision Transformer as a viable alternative to CNN. Different from convolutional networks, transformers operate with images as a sequence of patches, selectively attending to different image parts based on context. In addition, the attention mechanism in the transformer solves the problem with a small flame, thereby provide detection fire in the early stage. Since transformers using global self-attention, which conducts complex computing, we utilize fine-tuned Swin Transformer as our backbone architecture that computes self-attention with local windows. Thus, solving the classification problems with high-resolution images. Experimental results conducted on the image fire dataset demonstrate the promising capability of the model compared to state-of-the-art methods. Specifically, Vision Transformer obtains a classification accuracy of 98.54% on the publicly available dataset.
引用
收藏
页码:41 / 53
页数:13
相关论文
共 50 条
  • [1] Fire detection using vision transformer on power plant
    Zhang, Kaidi
    Wang, Binjun
    Tong, Xin
    Liu, Keke
    ENERGY REPORTS, 2022, 8 : 657 - 664
  • [2] Multiscale fire image detection method based on CNN and Transformer
    Wu, Shengbao
    Sheng, Buyun
    Fu, Gaocai
    Zhang, Daode
    Jian, Yuchao
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 49787 - 49811
  • [3] An intelligent fire detection approach through cameras based on computer vision methods
    Wu, Hao
    Wu, Deyang
    Zhao, Jinsong
    PROCESS SAFETY AND ENVIRONMENTAL PROTECTION, 2019, 127 : 245 - 256
  • [4] A modified vision transformer architecture with scratch learning capabilities for effective fire detection
    Yar, Hikmat
    Khan, Zulfiqar Ahmad
    Hussain, Tanveer
    Baik, Sung Wook
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 252
  • [5] A Vision-based Approach to Fire Detection
    Gomes, Pedro
    Santana, Pedro
    Barata, Jose
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2014, 11
  • [6] A novel approach based on convolutional neural networks ensemble for fire detection
    Belarbi, Farah
    Hassini, Abdelatif
    Benamara, Nadir Kamel
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (12) : 8805 - 8818
  • [7] Development and evaluation of a vision-based transfer learning approach for indoor fire and smoke detection
    Pincott, James
    Tien, Paige Wenbin
    Wei, Shuangyu
    Kaiser Calautit, John
    BUILDING SERVICES ENGINEERING RESEARCH & TECHNOLOGY, 2022, 43 (03) : 319 - 332
  • [8] Transformer-Based Fire Detection in Videos
    Mardani, Konstantina
    Vretos, Nicholas
    Daras, Petros
    SENSORS, 2023, 23 (06)
  • [9] Intelligent and vision-based fire detection systems: A survey
    Bu, Fengju
    Gharajeh, Mohammad Samadi
    IMAGE AND VISION COMPUTING, 2019, 91
  • [10] A Probabilistic Approach for Vision-Based Fire Detection in Videos
    Borges, Paulo Vinicius Koerich
    Izquierdo, Ebroul
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2010, 20 (05) : 721 - 731