Focal segmentation for robust 6D object pose estimation

被引:1
作者
Ye, Yuning [1 ]
Park, Hanhoon [1 ,2 ]
机构
[1] Pukyong Natl Univ, Grad Sch, Dept Artificial Intelligence Convergence, Pusan 48513, South Korea
[2] Pukyong Natl Univ, Div Elect & Commun Engn, Pusan 48513, South Korea
基金
新加坡国家研究基金会;
关键词
Object pose estimation; Focal segmentation; Keypoint detection; Severe occlusion; Deep learning; ACCURATE; NETWORK;
D O I
10.1007/s11042-023-16937-y
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the field of augmented reality, 6D pose estimation of rigid objects poses limitations and challenges. Most of the previous 6D pose estimation methods have trained deep neural networks to directly regress poses from input images or predict the 2D locations of 3D keypoints for pose estimation; thus, they are vulnerable to large occlusion. This study addresses the challenge of 6D pose estimation from a single RGB image under severe occlusion. A novel method is proposed that is based on PVNet but improves its performance. Similar to PVNet, our method regresses target object segments and pixel-wise direction vectors from an RGB image. Subsequently, the 2D locations of 3D keypoints are computed using the direction vectors of object pixels, and the 6D object pose is obtained using a PnP algorithm. However, accurate segmentation of object pixels is difficult, particularly under severe occlusion. To this end, a focal segmentation mechanism is proposed that ensures accurate complete segmentation of occluded objects. Extensive experiments on LINEMOD, LINEMOD-Occlusion datasets validate the effectiveness and superiority of our method. Our method improves the accuracy of PVNet by 1.09 and 5.14 on average in terms of the 2D reprojection error and ADD metric, respectively, without increasing the computational time.
引用
收藏
页码:47563 / 47585
页数:23
相关论文
共 39 条
[1]  
[Anonymous], About us
[2]   Uncertainty-Driven 6D Pose Estimation of Objects and Scenes from a Single RGB Image [J].
Brachmann, Eric ;
Michel, Frank ;
Krull, Alexander ;
Yang, Michael Ying ;
Gumhold, Stefan ;
Rother, Carsten .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3364-3372
[3]  
Brachmann E, 2014, LECT NOTES COMPUT SC, V8690, P536, DOI 10.1007/978-3-319-10605-2_35
[4]  
Do TT, 2018, Arxiv, DOI [arXiv:1802.10367, 10.48550/ARXIV.1802.10367]
[5]   Recovering 6D Object Pose and Predicting Next-Best-View in the Crowd [J].
Doumanoglou, Andreas ;
Kouskouridas, Rigas ;
Malassiotis, Sotiris ;
Kim, Tae-Kyun .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3583-3592
[6]   Real-time visual tracking of complex structures [J].
Drummond, T ;
Cipolla, R .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2002, 24 (07) :932-946
[7]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[8]   Efficient Multi-Person Hierarchical 3D Pose Estimation for Autonomous Driving [J].
Gu, Renshu ;
Wang, Gaoang ;
Hwang, Jenq-Neng .
2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, :163-168
[9]  
Hachiuma R, 2016, 2016 IEEE 2ND WORKSHOP ON EVERYDAY VIRTUAL REALITY (WEVR), P32, DOI 10.1109/WEVR.2016.7859541
[10]   SurfEmb: Dense and Continuous Correspondence Distributions for Object Pose Estimation with Learnt Surface Embeddings [J].
Haugaard, Rasmus Laurvig ;
Buch, Anders Glent .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :6739-6748