Multi Task-Guided 6D Object Pose Estimation

被引:0
作者
Thu-Uyen Nguyen [1 ]
Van-Duc Vu [1 ]
Van-Thiep Nguyen [1 ]
Ngoc-Anh Hoang [1 ]
Duy-Quang Vu [1 ]
Duc-Thanh Tran [1 ]
Khanh-Toan Phan [1 ]
Anh-Truong Mai [1 ]
Van-Hiep Duong [1 ]
Cong-Trinh Chan [1 ]
Ngoc-Trung Ho [1 ]
Quang-Tri Duong [1 ]
Phuc-Quan Ngo [1 ]
Dinh-Cuong Hoang [1 ]
机构
[1] FPT Univ Hanoi, Hanoi, Vietnam
来源
PROCEEDINGS OF THE 2024 9TH INTERNATIONAL CONFERENCE ON INTELLIGENT INFORMATION TECHNOLOGY, ICIIT 2024 | 2024年
关键词
Pose estimation; robot vision systems; intelligent systems; deep learning; supervised learning; machine vision;
D O I
10.1145/3654522.3654576
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Object pose estimation remains a fundamental challenge in computer vision, with cutting-edge methods relying on both RGB and depth data. Depth information is pivotal, offering crucial geometric cues that enable algorithms to navigate occlusions, fostering a more comprehensive scene under-standing and precise pose estimation. However, RGBD-based methods often require specialized depth sensors, which can be costlier and less accessible compared to standard RGB cameras. Consequently, research has explored techniques aiming to estimate object pose solely from color images. Yet, the absence of depth cues poses challenges in handling occlusions, comprehending object geometry, and resolving ambiguities arising from similar colors or textures. This paper introduces a end-to-end multi-task-guided object pose estimation method, utilizing RGB images as input and producing the 6D pose of multiple object instances. While our approach employs both depth and color images during training, inference relies solely on color images. We incorporate depth images to supervise a depth estimation branch, generating depth-aware features further refined through a cross-task attention module. These enhanced features are pivotal for our object pose estimation. Our method's innovation lies in significantly enhancing feature discriminability and robustness for object pose estimation. Through extensive experiments, we demonstrate competitive performance compared to state-of-the-art methods in object pose estimation.
引用
收藏
页码:215 / 222
页数:8
相关论文
共 50 条
  • [1] ConvPoseCNN: Dense Convolutional 6D Object Pose Estimation
    Capellen, Catherine
    Schwarz, Max
    Behnke, Sven
    PROCEEDINGS OF THE 15TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VOL 5: VISAPP, 2020, : 162 - 172
  • [2] Graph neural network for 6D object pose estimation
    Yin, Pengshuai
    Ye, Jiayong
    Lin, Guoshen
    Wu, Qingyao
    KNOWLEDGE-BASED SYSTEMS, 2021, 218
  • [3] Survey on 6D Pose Estimation of Rigid Object
    Chen, Jiale
    Zhang, Lijun
    Liu, Yi
    Xu, Chi
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 7440 - 7445
  • [4] Selective Embedding with Gated Fusion for 6D Object Pose Estimation
    Sun, Shantong
    Liu, Rongke
    Du, Qiuchen
    Sun, Shuqiao
    NEURAL PROCESSING LETTERS, 2020, 51 (03) : 2417 - 2436
  • [5] Temporally Consistent Object 6D Pose Estimation for Robot Control
    Zorina, Kateryna
    Priban, Vojtech
    Fourmy, Mederic
    Sivic, Josef
    Petrik, Vladimir
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (01): : 56 - 63
  • [6] Selective Embedding with Gated Fusion for 6D Object Pose Estimation
    Shantong Sun
    Rongke Liu
    Qiuchen Du
    Shuqiao Sun
    Neural Processing Letters, 2020, 51 : 2417 - 2436
  • [7] Focal segmentation for robust 6D object pose estimation
    Yuning Ye
    Hanhoon Park
    Multimedia Tools and Applications, 2024, 83 : 47563 - 47585
  • [8] Fundamental Coordinate Space for Object 6D Pose Estimation
    Wan, Boyan
    Zhang, Chen
    IEEE ACCESS, 2024, 12 : 146430 - 146440
  • [9] Focal segmentation for robust 6D object pose estimation
    Ye, Yuning
    Park, Hanhoon
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (16) : 47563 - 47585
  • [10] Confidence-Based 6D Object Pose Estimation
    Huang, Wei-Lun
    Hung, Chun-Yi
    Lin, I-Chen
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 3025 - 3035