EPFA-Net: An Enhanced Partial Feature Aggregation Network for Remote Sensing Object Detection

被引:0
作者
Wu, Zhen [1 ]
Zhang, Li [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
来源
PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024 | 2024年
关键词
Remote Sensing Object detection; Computer Vision; Convolutional Neural Network; Self-attention Mechanism;
D O I
10.1109/CSCWD61410.2024.10580866
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Deep learning models, including convolution-based and self-attention-based models, have been highly successful applied to remote sensing object detection owing to their powerful performance. However, convolution-based models are constrained by the locality of convolution, thereby limiting their detection performance; self-attention-based methods require a large number of training samples to achieve good performance and have low computational efficiency, which hinders their application in real-time detection tasks. To address these issues, we propose an enhanced partial feature aggregation network (EPFA-Net) for remote sensing object detection. In EPFA-Net, we design an enhanced partial feature aggregation (EPFA) module that consists of two parts. The first part is an efficient layer aggregation block that is used to extract local features. The second part adopts a partial feature aggregation (PFA) structure, enhancing gradient efficiency, and a shift-window attention block to extract global features, establishing long-range dependencies. Extensive experiments are conducted on two challenging remote sensing object detection datasets. Experimental results indicate that EPFA-Net is superior to four mainstream methods in detection performance. In addition, findings from ablation experiments suggest that EPFA-Net is efficient in inference speed because of the utilization of the PFA structure.
引用
收藏
页码:1116 / 1121
页数:6
相关论文
共 22 条
[1]  
[Anonymous], 2020, PROCEEDINGS OF THE I, DOI DOI 10.1109/CVPR42600.2020.01079
[2]   Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images [J].
Cheng, Gong ;
Zhou, Peicheng ;
Han, Junwei .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2016, 54 (12) :7405-7415
[3]  
Dosovitskiy A., 2020, INT C LEARN REPR, P1
[4]  
Gao L., 2023, IEEE T GEOSC REM SEN, V61
[5]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587
[6]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[7]   Densely Connected Convolutional Networks [J].
Huang, Gao ;
Liu, Zhuang ;
van der Maaten, Laurens ;
Weinberger, Kilian Q. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2261-2269
[8]  
Joseph RK, 2016, CRIT POL ECON S ASIA, P1
[9]   Rotation-Insensitive and Context-Augmented Object Detection in Remote Sensing Images [J].
Li, Ke ;
Cheng, Gong ;
Bu, Shuhui ;
You, Xiong .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2018, 56 (04) :2337-2348
[10]  
Li S., 2022, IEEE T GEOSC REM SEN, V60