Camouflaged Object Detection Based on Deep Learning with Attention-Guided Edge Detection and Multi-Scale Context Fusion

被引:1
|
作者
Wen, Yalin [1 ]
Ke, Wei [1 ,2 ]
Sheng, Hao [1 ,3 ,4 ]
机构
[1] Macao Polytech Univ, Fac Appl Sci, Macau 999078, Peoples R China
[2] Macao Polytech Univ, Engn Res Ctr Appl Technol Machine Translat & Artif, Minist Educ, Macau 999078, Peoples R China
[3] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[4] Beihang Univ, Zhongfa Aviat Inst, Hangzhou 310000, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 06期
关键词
camouflaged object detection; EfficientNet; salient object detection; deep learning; NETWORK;
D O I
10.3390/app14062494
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In nature, objects that use camouflage have features like colors and textures that closely resemble their background. This creates visual illusions that help them hide and protect themselves from predators. This similarity also makes the task of detecting camouflaged objects very challenging. Methods for camouflaged object detection (COD), which rely on deep neural networks, are increasingly gaining attention. These methods focus on improving model performance and computational efficiency by extracting edge information and using multi-layer feature fusion. Our improvement is based on researching ways to enhance efficiency in the encode-decode process. We have developed a variant model that combines Swin Transformer (Swin-T) and EfficientNet-B7. This model integrates the strengths of both Swin-T and EfficientNet-B7, and it employs an attention-guided tracking module to efficiently extract edge information and identify objects in camouflaged environments. Additionally, we have incorporated dense skip links to enhance the aggregation of deep-level feature information. A boundary-aware attention module has been incorporated into the final layer of the initial shallow information recognition phase. This module utilizes the Fourier transform to quickly relay specific edge information from the initially obtained shallow semantics to subsequent stages, thereby more effectively achieving feature recognition and edge extraction. In the latter phase, which is focused on deep semantic extraction, we employ a dense skip joint attention module to enhance the decoder's performance and efficiency, ensuring accurate capture of deep-level information, feature recognition, and edge extraction. In the later stage of deep semantic extraction, we use a dense skip joint attention module to improve the decoder's performance and efficiency in capturing precise deep information. This module efficiently identifies the specifics and edge information of undetected camouflaged objects across channels and spaces. Differing from previous methods, we introduce an adaptive pixel strength loss function for handling key captured information. Our proposed method shows strong competitive performance on three current benchmark datasets (CHAMELEON, CAMO, COD10K). Compared to 26 previously proposed methods using 4 measurement metrics, our approach exhibits favorable competitiveness.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Attention-Guided Deep Neural Network With Multi-Scale Feature Fusion for Liver Vessel Segmentation
    Yan, Qingsen
    Wang, Bo
    Zhang, Wei
    Luo, Chuan
    Xu, Wei
    Xu, Zhengqing
    Zhang, Yanning
    Shi, Qinfeng
    Zhang, Liang
    You, Zheng
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (07) : 2629 - 2642
  • [22] Pyramid attention object detection network with multi-scale feature fusion
    Chen, Xiu
    Li, Yujie
    Nakatoh, Yoshihisa
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 104
  • [23] An attention-guided multi-scale fusion network for surgical instrument segmentation
    Song, Mengqiu
    Zhai, Chenxu
    Yang, Lei
    Liu, Yanhong
    Bian, Guibin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 102
  • [24] Small Object Detection using Multi-scale Feature Fusion and Attention
    Liu, Baokai
    Du, Shiqiang
    Li, Jiacheng
    Wang, Jianhua
    Liu, Wenjie
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 7246 - 7251
  • [25] Multi-Scale Context Aggregation Network with Attention-Guided for Crowd Counting
    Wang, Xin
    Lv, Rongrong
    Zhao, Yang
    Yang, Tangwen
    Ruan, Qiuqi
    PROCEEDINGS OF 2020 IEEE 15TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP 2020), 2020, : 240 - 245
  • [26] Supervised Multi-Scale Attention-Guided Ship Detection in Optical Remote Sensing Images
    Hu, Jianming
    Zhi, Xiyang
    Jiang, Shikai
    Tang, Hao
    Zhang, Wei
    Bruzzone, Lorenzo
    IEEE Transactions on Geoscience and Remote Sensing, 2022, 60
  • [27] 3D Object Detection Based on Attention and Multi-Scale Feature Fusion
    Liu, Minghui
    Ma, Jinming
    Zheng, Qiuping
    Liu, Yuchen
    Shi, Gang
    SENSORS, 2022, 22 (10)
  • [28] Supervised Multi-Scale Attention-Guided Ship Detection in Optical Remote Sensing Images
    Hu, Jianming
    Zhi, Xiyang
    Jiang, Shikai
    Tang, Hao
    Zhang, Wei
    Bruzzone, Lorenzo
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [29] Attention-guided multi-scale infrared real-time detection of pedestrian and vehicle
    Zhang Y.
    Ji K.
    He Z.
    Chen G.
    Hongwai yu Jiguang Gongcheng/Infrared and Laser Engineering, 2024, 53 (05):
  • [30] Spatial attention-guided deformable fusion network for salient object detection
    Yang, Aiping
    Liu, Yan
    Cheng, Simeng
    Cao, Jiale
    Ji, Zhong
    Pang, Yanwei
    MULTIMEDIA SYSTEMS, 2023, 29 (05) : 2563 - 2573