Visible and Infrared Image Fusion of Forest Fire Scenes Based on Generative Adversarial Networks with Multi-Classification and Multi-Level Constraints

被引:5
作者
Jin, Qi [1 ]
Tan, Sanqing [1 ]
Zhang, Gui [1 ]
Yang, Zhigao [1 ]
Wen, Yijun [1 ]
Xiao, Huashun [1 ]
Wu, Xin [1 ]
机构
[1] Cent South Univ Forestry & Technol, Coll Forestry, Changsha 410004, Peoples R China
基金
中国国家自然科学基金;
关键词
forest fire scenes; visible and infrared images; image fusion; deep learning; GAN; NEURAL-NETWORK; FRAMEWORK; DOMAIN;
D O I
10.3390/f14101952
中图分类号
S7 [林业];
学科分类号
0829 ; 0907 ;
摘要
Aimed at addressing deficiencies in existing image fusion methods, this paper proposed a multi-level and multi-classification generative adversarial network (GAN)-based method (MMGAN) for fusing visible and infrared images of forest fire scenes (the surroundings of firefighters), which solves the problem that GANs tend to ignore visible contrast ratio information and detailed infrared texture information. The study was based on real-time visible and infrared image data acquired by visible and infrared binocular cameras on forest firefighters' helmets. We improved the GAN by, on the one hand, splitting the input channels of the generator into gradient and contrast ratio paths, increasing the depth of convolutional layers, and improving the extraction capability of shallow networks. On the other hand, we designed a discriminator using a multi-classification constraint structure and trained it against the generator in a continuous and adversarial manner to supervise the generator, generating better-quality fused images. Our results indicated that compared to mainstream infrared and visible image fusion methods, including anisotropic diffusion fusion (ADF), guided filtering fusion (GFF), convolutional neural networks (CNN), FusionGAN, and dual-discriminator conditional GAN (DDcGAN), the MMGAN model was overall optimal and had the best visual effect when applied to image fusions of forest fire surroundings. Five of the six objective metrics were optimal, and one ranked second-to-optimal. The image fusion speed was more than five times faster than that of the other methods. The MMGAN model significantly improved the quality of fused images of forest fire scenes, preserved the contrast ratio information of visible images and the detailed texture information of infrared images of forest fire scenes, and could accurately reflect information on forest fire scene surroundings.
引用
收藏
页数:17
相关论文
共 29 条
[1]   Fusion of Infrared and Visible Sensor Images Based on Anisotropic Diffusion and Karhunen-Loeve Transform [J].
Bavirisetti, Durga Prasad ;
Dhuli, Ravindra .
IEEE SENSORS JOURNAL, 2016, 16 (01) :203-209
[2]   Speeded-Up Robust Features (SURF) [J].
Bay, Herbert ;
Ess, Andreas ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2008, 110 (03) :346-359
[3]   Computer vision techniques for forest fire perception [J].
Dios, J. R. Martinez-de ;
Arrue, B. C. ;
Ollero, A. ;
Merino, L. ;
Gomez-Rodriguez, F. .
IMAGE AND VISION COMPUTING, 2008, 26 (04) :550-562
[4]   A Multiscale Normalization Method of a Mixed-Effects Model for Monitoring Forest Fires Using Multi-Sensor Data [J].
Feng, Lanbo ;
Xiao, Huashun ;
Yang, Zhigao ;
Zhang, Gui .
SUSTAINABILITY, 2022, 14 (03)
[5]   The Property of Frequency Shift in 2D-FRFT Domain With Application to Image Encryption [J].
Gao, Lei ;
Qi, Lin ;
Guan, Ling .
IEEE SIGNAL PROCESSING LETTERS, 2021, 28 :185-189
[6]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[7]  
Hasnat A, 2017, International Journal of Virtual and Augmented Reality, V1, P72, DOI [10.4018/ijvar.2017010106, 10.4018/IJVAR.2017010106, DOI 10.4018/IJVAR.2017010106]
[8]   An image contrast enhancement algorithm for grayscale images using particle swarm optimization [J].
Kanmani, Madheswari ;
Narsimhan, Venkateswaran .
MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (18) :23371-23387
[9]   Multi-focus image fusion based on sparse feature matrix decomposition and morphological filtering [J].
Li, Hui ;
Li, Li ;
Zhang, Jixiang .
OPTICS COMMUNICATIONS, 2015, 342 :1-11
[10]   Multifocus image fusion using region segmentation and spatial frequency [J].
Li, Shutao ;
Yang, Bin .
IMAGE AND VISION COMPUTING, 2008, 26 (07) :971-979