DiffPoseNet: Direct Differentiable Camera Pose Estimation

被引:25
作者
Parameshwara, Chethan M. [1 ]
Hari, Gokul [1 ]
Fermuller, Cornelia [1 ]
Sanket, Nitin J. [1 ]
Aloimonos, Yiannis [1 ]
机构
[1] Univ Maryland, College Pk, MD 20742 USA
来源
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022) | 2022年
基金
美国国家科学基金会;
关键词
ROBUST ESTIMATION; FLOW; EGOMOTION;
D O I
10.1109/CVPR52688.2022.00672
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Current deep neural network approaches for camera pose estimation rely on scene structure for 3D motion estimation, but this decreases the robustness and thereby makes cross-dataset generalization difficult. In contrast, classical approaches to structure from motion estimate 3D motion utilizing optical flow and then compute depth. Their accuracy, however, depends strongly on the quality of the optical flow To avoid this issue, direct methods have been proposed, which separate 3D motion from depth estimation, but compute 3D motion using only image gradients in the form of normal flow In this paper, we introduce a network NFlowNet, for normal flow estimation which is used to enforce robust and direct constraints. In particular, normal flow is used to estimate relative camera pose based on the cheirality (depth positivity) constraint. We achieve this by formulating the optimization problem as a differentiable cheirality layer, which allows for end-to-end learning of camera pose. We perform extensive qualitative and quantitative evaluation of the proposed DiffPoseNet's sensitivity to noise and its generalization across datasets. We compare our approach to existing state-of-the-art methods on KITTI, TartanAir, and TUM-RGBD datasets.
引用
收藏
页码:6835 / 6844
页数:10
相关论文
共 55 条
[1]   ESTIMATING THE HEADING DIRECTION USING NORMAL FLOW [J].
ALOIMONOS, Y ;
DURIC, Z .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 1994, 13 (01) :33-56
[2]  
Anand Adithya Prem, 2020, 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA), P1289, DOI 10.1109/ICMLA51294.2020.00202
[3]  
[Anonymous], 2017, PROC INT C MACH LEAR
[4]   Joint direct estimation of 3D geometry and 3D motion using spatio temporal gradients [J].
Barranco, Francisco ;
Fermueller, Cornelia ;
Aloimonos, Yiannis ;
Ros, Eduardo .
PATTERN RECOGNITION, 2021, 113 (113)
[5]   The robust estimation of multiple motions: Parametric and piecewise-smooth flow fields [J].
Black, MJ ;
Anandan, P .
COMPUTER VISION AND IMAGE UNDERSTANDING, 1996, 63 (01) :75-104
[6]   Self-calibration from image derivatives [J].
Brodsky, T ;
Fermüller, C .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2002, 48 (02) :91-114
[7]  
Brodsky Tomas., 1999, Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149), V2, P2146
[8]   High accuracy optical flow estimation based on a theory for warping [J].
Brox, T ;
Bruhn, A ;
Papenberg, N ;
Weickert, J .
COMPUTER VISION - ECCV 2004, PT 4, 2004, 2034 :25-36
[9]   Large Displacement Optical Flow: Descriptor Matching in Variational Motion Estimation [J].
Brox, Thomas ;
Malik, Jitendra .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2011, 33 (03) :500-513
[10]   A LIMITED MEMORY ALGORITHM FOR BOUND CONSTRAINED OPTIMIZATION [J].
BYRD, RH ;
LU, PH ;
NOCEDAL, J ;
ZHU, CY .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 1995, 16 (05) :1190-1208