RAOD: refined oriented detector with augmented feature in remote sensing images object detection

被引:6
作者
Shi, Qin [1 ]
Zhu, Yu [1 ]
Fang, Chuantao [1 ]
Wang, Nan [1 ]
Lin, Jiajun [1 ]
机构
[1] East China Univ Sci & Technol, Sch Informat Sci & Engn, Shanghai 200237, Peoples R China
关键词
Remote sensing image; Oriented object detection; Augmented feature pyramid; Deformable RoI pooling; Rotated RoI align;
D O I
10.1007/s10489-022-03393-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Object detection is a challenging task in remote sensing. Aerial images are distinguished by complex backgrounds, arbitrary orientations, and dense distributions. Considering those difficulties, this paper proposes a two-stage refined oriented detector with augmented features named RAOD. First, a novel Augmented Feature Pyramid Network (A-FPN) is built to enhance fusion both in spatial and channel dimensions. Specifically, it mainly consists of three modules: Scale Transfer Module (STM), Feature Aggregate Module (FAM) and Feature Refinement Module (FRM). STM reduces information loss when fusing features in the top-down pathway. FAM aggregates features from different scales. FRM aims to refine the integrated features using a lightweight attention module. Then, we adopt a two-step processing, which consists of a coarse stage and a refinement stage. In the coarse stage, deformable RoI pooling is adopted to improve the network's ability of modeling spatial transformations and then horizontal proposals are transformed into oriented ones. In the refinement stage, Rotated RoI align (RRoI align) is used to extract rotation-invariant features from rotated RoIs and further optimize the localization. To enhance stability and robustness during training, smooth Ln is chosen as regression loss as it has better ability in terms of robustness and stability than smooth L-1 loss. Extensive experiments on several rotation detection datasets demonstrate the effectiveness of our method. Results show that our method is able to achieve 79.78%, 74.7% and 94.82% on DOTA-v1.0, DOTA-v1.5 and HRSC2016, respectively.
引用
收藏
页码:15278 / 15294
页数:17
相关论文
共 59 条
[51]   Arbitrary-Oriented Object Detection with Circular Smooth Label [J].
Yang, Xue ;
Yan, Junchi .
COMPUTER VISION - ECCV 2020, PT VIII, 2020, 12353 :677-694
[52]   SCRDet: Towards More Robust Detection for Small, Cluttered and Rotated Objects [J].
Yang, Xue ;
Yang, Jirui ;
Yan, Junchi ;
Zhang, Yue ;
Zhang, Tengfei ;
Guo, Zhi ;
Sun, Xian ;
Fu, Kun .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8231-8240
[53]   Oriented Object Detection in Aerial Images with Box Boundary-Aware Vectors [J].
Yi, Jingru ;
Wu, Pengxiang ;
Liu, Bo ;
Huang, Qiaoying ;
Qu, Hui ;
Metaxas, Dimitris .
2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, :2149-2158
[54]   Feature Pyramid Transformer [J].
Zhang, Dong ;
Zhang, Hanwang ;
Tang, Jinhui ;
Wang, Meng ;
Hua, Xiansheng ;
Sun, Qianru .
COMPUTER VISION - ECCV 2020, PT XXVIII, 2020, 12373 :323-339
[55]   CAD-Net: A Context-Aware Detection Network for Objects in Remote Sensing Imagery [J].
Zhang, Gongjie ;
Lu, Shijian ;
Zhang, Wei .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2019, 57 (12) :10015-10024
[56]   ROSD: Refined Oriented Staged Detector for Object Detection in Aerial Image [J].
Zhang, Kun ;
Zeng, Qunqi ;
Yu, Xiangyu .
IEEE ACCESS, 2021, 9 :66560-66569
[57]   Toward Arbitrary-Oriented Ship Detection With Rotated Region Proposal and Discrimination Networks [J].
Zhang, Zenghui ;
Guo, Weiwei ;
Zhu, Shengnan ;
Yu, Wenxian .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2018, 15 (11) :1745-1749
[58]   Oriented Response Networks [J].
Zhou, Yanzhao ;
Ye, Qixiang ;
Qiu, Qiang ;
Jiao, Jianbin .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4961-4970
[59]   Spatial hierarchy perception and hard samples metric learning for high-resolution remote sensing image object detection [J].
Zhu, Dongjun ;
Xia, Shixiong ;
Zhao, Jiaqi ;
Zhou, Yong ;
Niu, Qiang ;
Yao, Rui ;
Chen, Ying .
APPLIED INTELLIGENCE, 2022, 52 (03) :3193-3208