A Multi-Scale Spatial Attention Region Proposal Network for High-Resolution Optical Remote Sensing Imagery

被引:5
作者
Dong, Ruchan [1 ,2 ]
Jiao, Licheng [3 ]
Zhang, Yan [1 ]
Zhao, Jin [4 ]
Shen, Weiyan [1 ]
机构
[1] Jinling Inst Technol, Nanjing 211169, Peoples R China
[2] Software Testing Engn Lab Jiangsu Prov, Nanjing 211169, Peoples R China
[3] Xidian Univ, Key Lab Intelligent Percept & Image Understanding, Minist Educ China, Xian 710071, Peoples R China
[4] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Peoples R China
关键词
high-resolution optical remote sensing images; multi-scale; spatial attention; region proposal;
D O I
10.3390/rs13173362
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Deep convolutional neural networks (DCNNs) are driving progress in object detection of high-resolution remote sensing images. Region proposal generation, as one of the key steps in object detection, has also become the focus of research. High-resolution remote sensing images usually contain various sizes of objects and complex background, small objects are easy to miss or be mis-identified in object detection. If the recall rate of region proposal of small objects and multi-scale objects can be improved, it will bring an improvement on the performance of the accuracy in object detection. Spatial attention is the ability to focus on local features in images and can improve the learning efficiency of DCNNs. This study proposes a multi-scale spatial attention region proposal network (MSA-RPN) for high-resolution optical remote sensing imagery. The MSA-RPN is an end-to-end deep learning network with a backbone network of ResNet. It deploys three novel modules to fulfill its task. First, the Scale-specific Feature Gate (SFG) focuses on features of objects by processing multi-scale features extracted from the backbone network. Second, the spatial attention-guided model (SAGM) obtains spatial information of objects from the multi-scale attention maps. Third, the Selective Strong Attention Maps Model (SSAMM) adaptively selects sliding windows according to the loss values from the system's feedback, and sends the windowed samples to the spatial attention decoder. Finally, the candidate regions and their corresponding confidences can be obtained. We evaluate the proposed network in a public dataset LEVIR and compare with several state-of-the-art methods. The proposed MSA-RPN yields a higher recall rate of region proposal generation, especially for small targets in remote sensing images.
引用
收藏
页数:13
相关论文
共 31 条
[1]  
Alexe B, 2010, PROC CVPR IEEE, P73, DOI 10.1109/CVPR.2010.5540226
[2]  
Almahairi A, 2016, PR MACH LEARN RES, V48
[3]   BING: Binarized Normed Gradients for Objectness Estimation at 300fps [J].
Cheng, Ming-Ming ;
Zhang, Ziming ;
Lin, Wen-Yan ;
Torr, Philip .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :3286-3293
[4]   Multi-scale object detection in remote sensing imagery with convolutional neural networks [J].
Deng, Zhipeng ;
Sun, Hao ;
Zhou, Shilin ;
Zhao, Juanping ;
Lei, Lin ;
Zou, Huanxin .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2018, 145 :3-22
[5]   Sig-NMS-Based Faster R-CNN Combining Transfer Learning for Small Target Detection in VHR Optical Remote Sensing Imagery [J].
Dong, Ruchan ;
Xu, Dazhuan ;
Zhao, Jin ;
Jiao, Licheng ;
An, Jungang .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2019, 57 (11) :8534-8545
[6]  
Gidaris S., 2016, BMVC, P1
[7]  
Gu CH, 2009, PROC CVPR IEEE, P1030, DOI 10.1109/CVPRW.2009.5206727
[8]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[9]   Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2015, 37 (09) :1904-1916
[10]   What Makes for Effective Detection Proposals? [J].
Hosang, Jan ;
Benenson, Rodrigo ;
Dollar, Piotr ;
Schiele, Bernt .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (04) :814-830