Denoising Diffusion Implicit Model for Camouflaged Object Detection

被引:0
作者
Cai, Wei [1 ]
Gao, Weijie [1 ]
Jiang, Xinhao [2 ]
Wang, Xin [1 ]
Di, Xingyu [1 ]
机构
[1] Xian Res Inst High Technol, Xian 710064, Peoples R China
[2] High Tech Inst, Fan Gong-Ting South St 12th, Qingzhou 262500, Peoples R China
关键词
camouflaged object detection; diffusion model; computer vision; feature fusion; location refinement;
D O I
10.3390/electronics13183690
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Camouflaged object detection (COD) is a challenging task that involves identifying objects that closely resemble their background. In order to detect camouflaged objects more accurately, we propose a diffusion model for the COD network called DMNet. DMNet formulates COD as a denoising diffusion process from noisy boxes to prediction boxes. During the training stage, random boxes diffuse from ground-truth boxes, and DMNet learns to reverse this process. In the sampling stage, DMNet progressively refines random boxes to prediction boxes. In addition, due to the camouflaged object's blurred appearance and the low contrast between it and the background, the feature extraction stage of the network is challenging. Firstly, we proposed a parallel fusion module (PFM) to enhance the information extracted from the backbone. Then, we designed a progressive feature pyramid network (PFPN) for feature fusion, in which the upsample adaptive spatial fusion module (UAF) balances the different feature information by assigning weights to different layers. Finally, a location refinement module (LRM) is constructed to make DMNet pay attention to the boundary details. We compared DMNet with other classical object-detection models on the COD10K dataset. Experimental results indicated that DMNet outperformed others, achieving optimal effects across six evaluation metrics and significantly enhancing detection accuracy.
引用
收藏
页数:21
相关论文
共 38 条
[1]  
Avrahami O, 2022, Arxiv, DOI [arXiv:2111.14818, DOI 10.48550/ARXIV.2111.14818]
[2]   YOLACT Real-time Instance Segmentation [J].
Bolya, Daniel ;
Zhou, Chong ;
Xiao, Fanyi ;
Lee, Yong Jae .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9156-9165
[3]   Cascade R-CNN: Delving into High Quality Object Detection [J].
Cai, Zhaowei ;
Vasconcelos, Nuno .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6154-6162
[4]  
Chung HYJ, 2022, Arxiv, DOI arXiv:2206.00941
[5]  
Ding X., 2019, arXiv
[6]   CenterNet: Keypoint Triplets for Object Detection [J].
Duan, Kaiwen ;
Bai, Song ;
Xie, Lingxi ;
Qi, Honggang ;
Huang, Qingming ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6568-6577
[7]   Structure and Content-Guided Video Synthesis with Diffusion Models [J].
Esser, Patrick ;
Chiu, Johnathan ;
Atighehchian, Parmida ;
Granskog, Jonathan ;
Germanidis, Anastasis .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, :7312-7322
[8]   Camouflaged Object Detection [J].
Fan, Deng-Ping ;
Ji, Ge-Peng ;
Sun, Guolei ;
Cheng, Ming-Ming ;
Shen, Jianbing ;
Shao, Ling .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :2774-2784
[9]   Object Detection with Discriminatively Trained Part-Based Models [J].
Felzenszwalb, Pedro F. ;
Girshick, Ross B. ;
McAllester, David ;
Ramanan, Deva .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2010, 32 (09) :1627-1645
[10]   TOOD: Task-aligned One-stage Object Detection [J].
Feng, Chengjian ;
Zhong, Yujie ;
Gao, Yu ;
Scott, Matthew R. ;
Huang, Weilin .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :3490-3499