Depth-Guided Progressive Network for Object Detection

被引:4
|
作者
Ma, Jia-Wei [1 ,2 ]
Liang, Min [2 ]
Chen, Song-Lu [1 ,2 ]
Chen, Feng [3 ]
Tian, Shu [2 ]
Qin, Jingyan [4 ]
Yin, Xu-Cheng [1 ,2 ]
机构
[1] Univ Sci & Technol Beijing, USTB EEasyTech Joint Lab Artificial Intelligence, Beijing 100083, Peoples R China
[2] Univ Sci & Technol Beijing, Dept Comp Sci & Technol, Beijing 100083, Peoples R China
[3] EEasy Technol Co Ltd, Zhuhai 519000, Peoples R China
[4] Univ Sci & Technol Beijing, Dept Ind Design, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Object detection; Detectors; Interference; Signal to noise ratio; Semantics; Location awareness; multi-scale object; depth-guided; progressive sampling;
D O I
10.1109/TITS.2022.3156365
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Multi-scale object detection in natural scenes is still challenging. To enhance the multi-scale perception capability, some algorithms combine the lower-level and higher-level information via multi-scale feature fusion strategies. However, the inherent spatial properties among instances and relations between foreground and background are ignored. In addition, the human-defined ``center-based'' regression quality evaluation strategy, predicting a high-to-low score based on a linear relationship with the distance to the center of ground-truth box, is not robust to scale-variant objects. In this work, we propose a Depth-Guided Progressive Network (DGPNet) for multi-scale object detection. Specifically, besides the prediction of classification and localization, the depth is estimated and used to guide the image features in a weighted manner to obtain a better spatial representation. Therefore, depth estimation and 2D object detection are simultaneously learned via a unified network, where the depth features are merged as auxiliary information into the detection branch to enhance the discrimination among multi-scale objects. Moreover, to overcome the difficulty of empirically fitting the localization quality function, high-quality predicted boxes on scale-variant objects are more adaptively obtained by an IoU-aware progressive sampling strategy. We divide the sampling process into two stages, i.e., ``statistical-aware'' and ``IoU-aware''. The former selects thresholds for positive samples based on statistical characteristics of multi-scale instances, and the latter further selects high-quality samples by IoU on the basis of the former. Therefore, the final ranking scores better reflect the quality of localization. Experiments verify that our method outperforms state-of-the-art methods on the KINS and Cityscapes dataset.
引用
收藏
页码:19523 / 19533
页数:11
相关论文
共 50 条
  • [41] DIFu: Depth-Guided Implicit Function for Clothed Human Reconstruction
    Song, Dae-Young
    Lee, HeeKyung
    Seo, Jeongil
    Cho, Donghyeon
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 8738 - 8747
  • [42] Novel View Synthesis via Depth-guided Skip Connections
    Hou, Yuxin
    Solin, Arno
    Kannala, Juho
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 3118 - 3127
  • [43] Depth-Guided NeRF Training via Earth Mover's Distance
    Rau, Anita
    Aklilu, Josiah
    Holsinger, F. Christopher
    Yeung-Levy, Serena
    COMPUTER VISION - ECCV 2024, PT LXIV, 2025, 15122 : 1 - 17
  • [44] Unsupervised Semantic Segmentation Through Depth-Guided Feature Correlation and Sampling
    Sick, Leon
    Engel, Dominik
    Hermosilla, Pedro
    Ropinski, Timo
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 3637 - 3646
  • [45] Progressive Self-Guided Loss for Salient Object Detection
    Yang, Sheng
    Lin, Weisi
    Lin, Guosheng
    Jiang, Qiuping
    Liu, Zichuan
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 : 8426 - 8438
  • [46] DGECN: A Depth-Guided Edge Convolutional Network for End-to-End 6D Pose Estimation
    Cao, Tuo
    Luo, Fei
    Fu, Yanping
    Zhang, Wenxiao
    Zheng, Shengjie
    Xiao, Chunxia
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 3773 - 3782
  • [47] Enhancing View Synthesis with Depth-Guided Neural Radiance Fields and Improved Depth Completion
    Wang, Bojun
    Zhang, Danhong
    Su, Yixin
    Zhang, Huajun
    SENSORS, 2024, 24 (06)
  • [48] Depth guided feature selection for RGBD salient object detection
    Li, Zun
    Lang, Congyan
    Li, Guanqin
    Wang, Tao
    Li, Yidong
    NEUROCOMPUTING, 2023, 519 : 57 - 68
  • [49] Guided residual network for RGB-D salient object detection with efficient depth feature learning
    Wang, Jian
    Chen, Shuhan
    Lv, Xiao
    Xu, Xiuqi
    Hu, Xuelong
    VISUAL COMPUTER, 2022, 38 (05): : 1803 - 1814
  • [50] Guided residual network for RGB-D salient object detection with efficient depth feature learning
    Jian Wang
    Shuhan Chen
    Xiao Lv
    Xiuqi Xu
    Xuelong Hu
    The Visual Computer, 2022, 38 : 1803 - 1814