Uncertainty-Driven Dense Two-View Structure From Motion

被引:6
|
作者
Chen, Weirong [1 ]
Kumar, Suryansh [1 ]
Yu, Fisher [1 ]
机构
[1] Swiss Fed Inst Technol, VIS Grp, CH-8092 Zurich, Switzerland
关键词
Cameras; Optical flow; Pipelines; Pose estimation; Reliability; Uncertainty; Neural networks; Dense structure from motion; uncertainty prediction; optical flow; weighted bundle adjustment;
D O I
10.1109/LRA.2023.3242153
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
This work introducesan effective and practical solution to the dense two-view structure from motion (SfM) problem. One vital question addressed is how to mindfully use per-pixel optical flow correspondence betweentwo frames for accurate pose estimation-as perfect per-pixel correspondence between two images is difficult, if not impossible, to establish. With the carefully estimated camera pose and predicted per-pixel optical flow correspondences, a dense depth of the scene is computed. Later, an iterative refinement procedure is introduced to further improve optical flow matching confidence, camera pose, and depth, exploiting their inherent dependency in rigid SfM. The fundamental idea presented is to benefit from per-pixel uncertainty in the optical flow estimation and provide robustness to the dense SfM system via an online refinement. Concretely, we introduce a pipeline consisting of (i) an uncertainty-aware dense optical flow estimation approach that provides per-pixel correspondence with their confidence score of matching; (ii) a weighted dense bundle adjustment formulation that depends on optical flow uncertainty and bidirectional optical flow consistency to refine both pose and depth; (iii) A depth estimation network that considers its consistency with the estimated poses and optical flow respecting epipolar constraint. Extensive experiments show that the proposed approach achieves remarkable depth accuracy and state-of-the-art camera pose results superseding SuperPoint and SuperGlue accuracy when tested on benchmark datasets such as DeMoN, YFCC100 M, and ScanNet.
引用
收藏
页码:1763 / 1770
页数:8
相关论文
共 50 条
  • [1] Two-view multibody structure from motion
    Vidal, Rene
    Ma, Yi
    Soatto, Stefano
    Sastry, Shankar
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2006, 68 (01) : 7 - 25
  • [2] Two-View Multibody Structure from Motion
    René Vidal
    Yi Ma
    Stefano Soatto
    Shankar Sastry
    International Journal of Computer Vision, 2006, 68 : 7 - 25
  • [3] Two-View Structure-from-Motion Algorithm
    Zhou, Haoyin
    Zhang, Tao
    2014 11TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), 2014, : 3345 - 3350
  • [4] Deep Two-View Structure-from-Motion Revisited
    Wang, Jianyuan
    Zhong, Yiran
    Dai, Yuchao
    Birchfield, Stan
    Zhang, Kaihao
    Smolyanskiy, Nikolai
    Li, Hongdong
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8949 - 8958
  • [5] Two-view multibody structure-and-motion with outliers
    Schindler, K
    Suter, D
    2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 2, PROCEEDINGS, 2005, : 676 - 683
  • [6] Uncertainty-Driven View Planning for Underwater Inspection
    Hollinger, Geoffrey A.
    Englot, Brendan
    Hover, Franz
    Mitra, Urbashi
    Sukhatme, Gaurav S.
    2012 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2012, : 4884 - 4891
  • [7] Two-View Structure-from-Motion with Multiple Feature Detector Operators
    Dippold, Elisabeth Johanna
    Tsai, Fuan
    REMOTE SENSING, 2023, 15 (03)
  • [8] Match Selection and Refinement for Highly Accurate Two-View Structure from Motion
    Liu, Zhe
    Monasse, Pascal
    Marlet, Renaud
    COMPUTER VISION - ECCV 2014, PT II, 2014, 8690 : 818 - 833
  • [9] DeepMLE: A Robust Deep Maximum Likelihood Estimator for Two-view Structure from Motion
    Xiao, Yuxi
    Li, Li
    Li, Xiaodi
    Yao, Jian
    2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 10643 - 10650
  • [10] Two-view motion segmentation from Linear Programming Relaxation
    Li, Hongdong
    2007 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-8, 2007, : 49 - 56