Improved YOLOv3 Based on Attention Mechanism for Fast and Accurate Ship Detection in Optical Remote Sensing Images

被引:89
作者
Chen, Liqiong [1 ]
Shi, Wenxuan [2 ]
Deng, Dexiang [1 ]
机构
[1] Wuhan Univ, Sch Elect Informat, Wuhan 430072, Peoples R China
[2] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金;
关键词
ship detection; optical remote sensing images; multi-class ship detection; dilated attention module; real-time speed; URBAN SPRAWL; SALIENCY; DATASET;
D O I
10.3390/rs13040660
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Ship detection is an important but challenging task in the field of computer vision, partially due to the minuscule ship objects in optical remote sensing images and the interference of clouds occlusion and strong waves. Most of the current ship detection methods focus on boosting detection accuracy while they may ignore the detection speed. However, it is also indispensable to increase ship detection speed because it can provide timely ocean rescue and maritime surveillance. To solve the above problems, we propose an improved YOLOv3 (ImYOLOv3) based on attention mechanism, aiming to achieve the best trade-off between detection accuracy and speed. First, to realize high-efficiency ship detection, we adopt the off-the-shelf YOLOv3 as our basic detection framework due to its fast speed. Second, to boost the performance of original YOLOv3 for small ships, we design a novel and lightweight dilated attention module (DAM) to extract discriminative features for ship targets, which can be easily embedded into the basic YOLOv3. The integrated attention mechanism can help our model learn to suppress irrelevant regions while highlighting salient features useful for ship detection task. Furthermore, we introduce a multi-class ship dataset (MSD) and explicitly set supervised subclass according to the scales and moving states of ships. Extensive experiments verify the effectiveness and robustness of ImYOLOv3, and show that our method can accurately detect ships with different scales in different backgrounds, while at a real-time speed.
引用
收藏
页码:1 / 18
页数:18
相关论文
共 69 条
[21]  
Kang M, 2017, 2017 INTERNATIONAL WORKSHOP ON REMOTE SENSING WITH INTELLIGENT PROCESSING (RSIP 2017)
[22]   Contextual Region-Based Convolutional Neural Network with Multilayer Fusion for SAR Ship Detection [J].
Kang, Miao ;
Ji, Kefeng ;
Leng, Xiangguang ;
Lin, Zhao .
REMOTE SENSING, 2017, 9 (08)
[23]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[24]   HSF-Net: Multiscale Deep Feature Embedding for Ship Detection in Optical Remote Sensing Imagery [J].
Li, Qingpeng ;
Mou, Lichao ;
Liu, Qingjie ;
Wang, Yunhong ;
Zhu, Xiao Xiang .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2018, 56 (12) :7147-7161
[25]   Selective Kernel Networks [J].
Li, Xiang ;
Wang, Wenhai ;
Hu, Xiaolin ;
Yang, Jian .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :510-519
[26]   Scale-Aware Trident Networks for Object Detection [J].
Li, Yanghao ;
Chen, Yuntao ;
Wang, Naiyan ;
Zhang, Zhaoxiang .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6053-6062
[27]  
Li ZM, 2018, LECT NOTES COMPUT SC, V11213, P339, DOI [10.1007/978-3-030-01240-3_21, 10.1007/978-3-030-01219-9_23]
[28]   Fully Convolutional Network With Task Partitioning for Inshore Ship Detection in Optical Remote Sensing Images [J].
Lin, Haoning ;
Shi, Zhenwei ;
Zou, Zhengxia .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2017, 14 (10) :1665-1669
[29]   Focal Loss for Dense Object Detection [J].
Lin, Tsung-Yi ;
Goyal, Priya ;
Girshick, Ross ;
He, Kaiming ;
Dollar, Piotr .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2999-3007
[30]   Feature Pyramid Networks for Object Detection [J].
Lin, Tsung-Yi ;
Dollar, Piotr ;
Girshick, Ross ;
He, Kaiming ;
Hariharan, Bharath ;
Belongie, Serge .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :936-944