Multi-Scale Feature Enhancement Method for Underwater Object Detection

被引:2
作者
Li, Mengpan [1 ]
Liu, Wenhao [1 ]
Shao, Changbin [1 ]
Qin, Bin [1 ]
Tian, Ali [2 ]
Yu, Hualong [1 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Comp Sci, Zhenjiang 212100, Peoples R China
[2] Jiangsu Univ Sci & Technol, Sch Naval Architecture & Ocean Engn, Zhenjiang 212100, Peoples R China
来源
SYMMETRY-BASEL | 2025年 / 17卷 / 01期
基金
中国国家自然科学基金;
关键词
underwater object detection; convolutional neural network; YOLO detection model; multi-scale feature enhancement; local awareness operation; feature pyramid network; RECOGNITION; LIGHTWEIGHT; YOLO;
D O I
10.3390/sym17010063
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
With deep-learning-based object detection methods reaching industrial-level performance, underwater object detection has emerged as a significant application. However, it is often challenged by dense small instances and image blurring due to the water medium. In this paper, a Multi-Scale Feature Enhancement(MSFE) method is presented to address the challenges triggered by water bodies. In brief, MSFE attempts to achieve dual multi-scale information integration through the internal structural design of the basic C2F module in the Backbone network and the external global design of the feature pyramid network (FPN). For the internal multi-scale implementation, a LABNK module is constructed to address the vanishing or weakening phenomenon of fine-grained features during feature extraction. Specifically, it adopts a symmetrical structure to collaboratively capture two types of local receptive field information. Furthermore, to enhance the information integration ability between inter-layer features in FPN, a shallow feature branch is injected to supplement detailed features for the subsequent integration of multi-scale features. This operation is mainly supported by the fact that large-sized features from the shallow layer usually carry rich, fine-grained information. Taking the typical YOLOv8n as the benchmark model, extensive experimental comparisons on public underwater datasets (DUO and RUOD) demonstrated the effectiveness of the presented MSFE method. For example, taking the rigorous mAP (50:95) as an evaluation metric, it can achieve an accuracy improvement of about 2.8%.
引用
收藏
页数:19
相关论文
共 69 条
[1]   Color-based underwater object recognition using water light attenuation [J].
Bazeille, Stephane ;
Quidu, Isabelle ;
Jaulin, Luc .
INTELLIGENT SERVICE ROBOTICS, 2012, 5 (02) :109-118
[2]  
Beijbom O, 2012, PROC CVPR IEEE, P1170, DOI 10.1109/CVPR.2012.6247798
[3]  
Bochkovskiy A, 2020, Arxiv, DOI [arXiv:2004.10934, DOI 10.48550/ARXIV.2004.10934]
[4]   Cascade R-CNN: Delving into High Quality Object Detection [J].
Cai, Zhaowei ;
Vasconcelos, Nuno .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6154-6162
[5]  
Chen L, 2024, Arxiv, DOI arXiv:2410.05577
[6]   Underwater object detection using Invert Multi-Class Adaboost with deep learning [J].
Chen, Long ;
Liu, Zhihua ;
Tong, Lei ;
Jiang, Zheheng ;
Wang, Shengke ;
Dong, Junyu ;
Zhou, Huiyu .
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
[7]  
Chen X, 2010, INT CONF SIGN PROCES, P793, DOI 10.1109/ICOSP.2010.5655926
[8]   Monocular Vision-Based Underwater Object Detection [J].
Chen, Zhe ;
Zhang, Zhen ;
Dai, Fengzhao ;
Bu, Yang ;
Wang, Huibin .
SENSORS, 2017, 17 (08)
[9]  
Cheng E., 2016, Int. J. Circuits Syst. Signal Process, V10, P291
[10]   Attention-Based Lightweight YOLOv8 Underwater Target Recognition Algorithm [J].
Cheng, Shun ;
Wang, Zhiqian ;
Liu, Shaojin ;
Han, Yan ;
Sun, Pengtao ;
Li, Jianrong .
SENSORS, 2024, 24 (23)