SEMI-DECOUPLED 6D POSE ESTIMATION VIA MULTI-MODAL FEATURE FUSION

被引:1
|
作者
Zhang, Zhenhu [1 ]
Cao, Xin [3 ]
Jin, Li [3 ]
Qin, Xueying [3 ]
Tong, Ruofeng [2 ]
机构
[1] Zhejiang Univ, Sch Software Technol, Hangzhou, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[3] Shandong Univ, Sch Software, Shandong, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024 | 2024年
基金
中国国家自然科学基金;
关键词
6D pose estimation; Multi-modal fusion; Semi-Decoupling;
D O I
10.1109/ICASSP48485.2024.10447649
中图分类号
学科分类号
摘要
The existing methods for 6D pose estimation based on RGBD employ RGB images and observed point cloud derived from depth maps as input, then concurrently predicting both rotation and translation. However, rotation and translation possess distinct characteristics and scale ranges, and their simultaneous prediction can lead to mutual influence in the network parameter space. Additionally, the observed point cloud are susceptible to systematic noise and partial data loss, presenting challenges for the network to capture comprehensive object features. To address these issues, we propose the Semi-Decoupled 6D pose estimation via multi-modal feature fusion (SD6D). SD6D comprises a Multi-Modal Fusion Module and a Semi-Decoupled Prediction Module. The former dynamically fuses different modal data (RGB, depth, CAD model) based on their inter-modality correlations, aiding in establishing 2D-3D correspondences and addressing issues stemming from systematic noise and partial data loss. The latter semi-decouples the prediction of rotation and translation, predicting them separately based on their distinct characteristics. We conducted experiments on two popular benchmark datasets, which prove the superiority of our method.
引用
收藏
页码:2610 / 2614
页数:5
相关论文
共 50 条
  • [1] Deep Fusion for Multi-Modal 6D Pose Estimation
    Lin, Shifeng
    Wang, Zunran
    Zhang, Shenghao
    Ling, Yonggen
    Yang, Chenguang
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (04) : 6540 - 6549
  • [2] A Transformer-based multi-modal fusion network for 6D pose estimation
    Hong, Jia-Xin
    Zhang, Hong-Bo
    Liu, Jing-Hua
    Lei, Qing
    Yang, Li-Jie
    Du, Ji-Xiang
    INFORMATION FUSION, 2024, 105
  • [3] Recovering 6D Object Pose: A Review and Multi-modal Analysis
    Sahin, Caner
    Kim, Tae-Kyun
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT VI, 2019, 11134 : 15 - 31
  • [4] Mitigating imbalances in heterogeneous feature fusion for multi-class 6D pose estimation
    Wang, Huafeng
    Zhang, Haodu
    Liu, Wanquan
    Lv, Weifeng
    Gu, Xianfeng
    Guo, Kexin
    KNOWLEDGE-BASED SYSTEMS, 2024, 297
  • [5] A modal fusion network with dual attention mechanism for 6D pose estimation
    Wei, Liangrui
    Xie, Feifei
    Sun, Lin
    Chen, Jinpeng
    Zhang, Zhipeng
    VISUAL COMPUTER, 2024, 40 (10): : 7411 - 7425
  • [6] 6D Pose Estimation with Correlation Fusion
    Cheng, Yi
    Zhu, Hongyuan
    Sun, Ying
    Acar, Cihan
    Jing, Wei
    Wu, Yan
    Li, Liyuan
    Tan, Cheston
    Lim, Joo-Hwee
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2988 - 2994
  • [7] An Improved Estimation Algorithm of Space Targets Pose Based on Multi-Modal Feature Fusion
    Hua, Jiang
    Hao, Tonglin
    Zeng, Liangcai
    Yu, Gui
    MATHEMATICS, 2021, 9 (17)
  • [8] A RGB-D feature fusion network for occluded object 6D pose estimation
    Song, Yiwei
    Tang, Chunhui
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (8-9) : 6309 - 6319
  • [9] 6D Object Pose Estimation Based on Cross-Modality Feature Fusion
    Jiang, Meng
    Zhang, Liming
    Wang, Xiaohua
    Li, Shuang
    Jiao, Yijie
    SENSORS, 2023, 23 (19)
  • [10] A Novel Depth and Color Feature Fusion Framework for 6D Object Pose Estimation
    Zhou, Guangliang
    Yan, Yi
    Wang, Deming
    Chen, Qijun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 : 1630 - 1639