A Multiobject Detection Scheme Based on Deep Learning for Infrared Images

被引:5
作者
Jiang, Chengyang [1 ]
Han, Jian-Jun [1 ]
机构
[1] Huazhong Univ Sci & Technol, Sch Comp Sci & Technol, Wuhan 430074, Hubei, Peoples R China
基金
美国国家科学基金会;
关键词
Detectors; Feature extraction; Object detection; Computational modeling; Superresolution; Image color analysis; Convolution; Multi-object detection; infrared image; anchor-based detection; attention mechanism; model pruning; OBJECT DETECTION; RECOGNITION; NETWORK;
D O I
10.1109/ACCESS.2022.3194037
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-object detection is one of essential prerequisites for autonomous driving under complicated traffic environments. Specially, for the cases of nighttime driving, the accuracy of object detector for those images captured by infrared cameras is inevitably affected by several unanticipated issues (such as dim and unexpected light sources), which may significantly degrade the performance of detecting multiple objects (e.g., pedestrians and vehicles). To address such problem, we propose a Light-Weighted Multi-Object Detection Scheme for Infrared Images (LW-MODS-IRI). First, we select an effective super-resolution (SR) method to enhance and repair those infrared images, which can facilitate restoring the details of images for better detection accuracy. Next, those enhanced images are trained by our LW-MODS-IRI scheme. In contrast with previous methods, we develop an adaptive computing policy for anchor clustering that aims at locating the bound of targets in a reasonable manner for higher detection precision. Moreover, a modified attention mechanism is exploited to enlarge the receptive fields for promoting the detection performance of those large objects. Furthermore, we propose an improved channel-level pruning technique to generate both compact and accurate model after the training with sparsity. Finally, the extensive experiment results on Teledyne FLIR ADAS dataset (FLIR) demonstrate that the LW-MODS-IRI scheme can have both higher inference accuracy (e.g., 10.7% more) and higher detection efficiency (e.g., 4.13 times faster) compared to the existing object detectors, which typically reveals its practical viability when it is applied to in-vehicle systems.
引用
收藏
页码:78939 / 78952
页数:14
相关论文
共 50 条
[1]   Multiple Hypothesis Detection and Tracking Using Deep Learning for Video Traffic Surveillance [J].
Ait Abdelali, Hamd ;
Derrouz, Hatim ;
Zennayi, Yahya ;
Haj Thami, Rachid Oulad ;
Bourzeix, Francois .
IEEE ACCESS, 2021, 9 :164282-164291
[2]   CNN-Based Object Recognition and Tracking System to Assist Visually Impaired People [J].
Ashiq, Fahad ;
Asif, Muhammad ;
Bin Ahmad, Maaz ;
Zafar, Sadia ;
Masood, Khalid ;
Mahmood, Toqeer ;
Mahmood, Muhammad Tariq ;
Lee, Ik Hyun .
IEEE ACCESS, 2022, 10 :14819-14834
[3]  
asia, TEL FLIR
[4]  
Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
[5]   Deformable Convolutional Networks [J].
Dai, Jifeng ;
Qi, Haozhi ;
Xiong, Yuwen ;
Li, Yi ;
Zhang, Guodong ;
Hu, Han ;
Wei, Yichen .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :764-773
[6]  
Denton E, 2014, ADV NEUR IN, V27
[7]   Learning a Deep Convolutional Network for Image Super-Resolution [J].
Dong, Chao ;
Loy, Chen Change ;
He, Kaiming ;
Tang, Xiaoou .
COMPUTER VISION - ECCV 2014, PT IV, 2014, 8692 :184-199
[8]  
Du Guoguang, 2019, Vision -based Robotic Grasping From Object Localization, Object Pose Estimation to Grasp Estimation for Parallel Grippers: A Review
[9]   Deep learning-based image recognition for autonomous driving [J].
Fujiyoshi, Hironobu ;
Hirakawa, Tsubasa ;
Yamashita, Takayoshi .
IATSS RESEARCH, 2019, 43 (04) :244-252
[10]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587