Swin Transformer-Based Object Detection Model Using Explainable Meta-Learning Mining

被引:9
作者
Baek, Ji-Won [1 ]
Chung, Kyungyong [2 ]
机构
[1] Kyonggi Univ, Dept Comp Sci, Suwon 16227, South Korea
[2] Kyonggi Univ, Div AI Comp Sci & Engn, Suwon 16227, South Korea
来源
APPLIED SCIENCES-BASEL | 2023年 / 13卷 / 05期
关键词
Swin Transformer; object detection; meta-learning; explainable AI; data mining; anomaly detection; DEEP; NETWORKS;
D O I
10.3390/app13053213
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In order to minimize damage in the event of a fire, the ignition point must be detected and dealt with before the fire spreads. However, the method of detecting fire by heat or fire is more damaging because it can be detected after the fire has spread. Therefore, this study proposes a Swin Transformer-based object detection model using explainable meta-learning mining. The proposed method merges the Swin Transformer and YOLOv3 model and applies meta-learning so as to build an explainable object detection model. In order for efficient learning with small data in the course of learning, it applies Few-Shot Learning. To find the causes of the object detection results, Grad-CAM as an explainable visualization method is used. It detects small objects of smoke in the fire image data and classifies them according to the color of the smoke generated when a fire breaks out. Accordingly, it is possible to predict and classify the risk of fire occurrence to minimize damage caused by fire. In this study, with the use of Mean Average Precision (mAP), performance evaluation is carried out in two ways. First, the performance of the proposed object detection model is evaluated. Secondly, the performance of the proposed method is compared with a conventional object detection method's performance. In addition, the accuracy comparison using the confusion matrix and the suitability of real-time object detection using FPS are judged. Given the results of the evaluation, the proposed method supports accurate and real-time monitoring and analysis.
引用
收藏
页数:14
相关论文
共 42 条
[1]  
AI Hub, About us
[2]  
[Anonymous], NATL FIRE INFORM SYS
[3]  
Arnab A., 2021, P IEEECVF INT C COMP
[4]   YOLACT plus plus Better Real-Time Instance Segmentation [J].
Bolya, Daniel ;
Zhou, Chong ;
Xiao, Fanyi ;
Lee, Yong Jae .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (02) :1108-1121
[5]   YOLACT Real-time Instance Segmentation [J].
Bolya, Daniel ;
Zhou, Chong ;
Xiao, Fanyi ;
Lee, Yong Jae .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9156-9165
[6]   A Multi-Class Multi-Movement Vehicle Counting Framework for Traffic Analysis in Complex Areas Using CCTV Systems [J].
Bui, Khac-Hoai Nam ;
Yi, Hongsuk ;
Cho, Jiho .
ENERGIES, 2020, 13 (08)
[7]   AP-Loss for Accurate One-Stage Object Detection [J].
Chen, Kean ;
Lin, Weiyao ;
Li, Jianguo ;
See, John ;
Wang, Ji ;
Zou, Junni .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (11) :3782-3798
[8]   A small attentional YOLO model for landslide detection from satellite remote sensing images [J].
Cheng, Libo ;
Li, Jia ;
Duan, Ping ;
Wang, Mingguo .
LANDSLIDES, 2021, 18 (08) :2751-2765
[9]   Dynamic Head: Unifying Object Detection Heads with Attentions [J].
Dai, Xiyang ;
Chen, Yinpeng ;
Xiao, Bin ;
Chen, Dongdong ;
Liu, Mengchen ;
Yuan, Lu ;
Zhang, Lei .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :7369-7378
[10]   YOLO-Former: Marrying YOLO and Transformer for Foreign Object Detection [J].
Dai, Yuan ;
Liu, Weiming ;
Wang, Heng ;
Xie, Wei ;
Long, Kejun .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71