Event-Based, 6-DOF Camera Tracking from Photometric Depth Maps

被引:129
作者
Gallego, Guillermo [1 ,2 ,3 ]
Lund, Jon E. A. [1 ,2 ,3 ]
Mueggler, Elias [1 ,2 ,3 ]
Rebecq, Henri [1 ,2 ,3 ]
Delbruck, Tobi [1 ,2 ,3 ]
Scaramuzza, Davide [1 ,2 ,3 ]
机构
[1] Univ Zurich, Robot & Percept Grp, Dept Informat, CH-8092 Zurich, Switzerland
[2] Univ Zurich, Dept Neuroinformat, CH-8092 Zurich, Switzerland
[3] Swiss Fed Inst Technol, CH-8092 Zurich, Switzerland
关键词
Event-based vision; pose tracking; dynamic vision sensor; Bayes filter; asynchronous processing; conjugate priors; low latency; high speed; AR/VR; VISUAL ODOMETRY; VISION; PIXEL;
D O I
10.1109/TPAMI.2017.2769655
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event cameras are bio-inspired vision sensors that output pixel-level brightness changes instead of standard intensity frames. These cameras do not suffer from motion blur and have a very high dynamic range, which enables them to provide reliable visual information during high-speed motions or in scenes characterized by high dynamic range. These features, along with a very low power consumption, make event cameras an ideal complement to standard cameras for VR/AR and video game applications. With these applications in mind, this paper tackles the problem of accurate, low-latency tracking of an event camera from an existing photometric depth map (i.e., intensity plus depth information) built via classic dense reconstruction pipelines. Our approach tracks the 6-DOF pose of the event camera upon the arrival of each event, thus virtually eliminating latency. We successfully evaluate the method in both indoor and outdoor scenes and show that-because of the technological advantages of the event camera-our pipeline works in scenes characterized by high-speed motion, which are still inaccessible to standard cameras.
引用
收藏
页码:2402 / 2412
页数:11
相关论文
共 35 条
  • [1] [Anonymous], 2016, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2016.102
  • [2] Belbachir AN, 2010, SMART CAMERAS, P1, DOI 10.1007/978-1-4419-0953-4
  • [3] Bishop Christopher M, 2016, Pattern recognition and machine learning
  • [4] A 240 x 180 130 dB 3 μs Latency Global Shutter Spatiotemporal Vision Sensor
    Brandli, Christian
    Berner, Raphael
    Yang, Minhao
    Liu, Shih-Chii
    Delbruck, Tobi
    [J]. IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2014, 49 (10) : 2333 - 2341
  • [5] Censi A, 2014, IEEE INT CONF ROBOT, P703, DOI 10.1109/ICRA.2014.6906931
  • [6] Cook M, 2011, 2011 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), P770, DOI 10.1109/IJCNN.2011.6033299
  • [7] LSD-SLAM: Large-Scale Direct Monocular SLAM
    Engel, Jakob
    Schoeps, Thomas
    Cremers, Daniel
    [J]. COMPUTER VISION - ECCV 2014, PT II, 2014, 8690 : 834 - 849
  • [8] SVO: Semidirect Visual Odometry for Monocular and Multicamera Systems
    Forster, Christian
    Zhang, Zichao
    Gassner, Michael
    Werlberger, Manuel
    Scaramuzza, Davide
    [J]. IEEE TRANSACTIONS ON ROBOTICS, 2017, 33 (02) : 249 - 265
  • [9] Forster C, 2014, IEEE INT CONF ROBOT, P15, DOI 10.1109/ICRA.2014.6906584
  • [10] Accurate Angular Velocity Estimation With an Event Camera
    Gallego, Guillermo
    Scaramuzza, Davide
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2017, 2 (02): : 632 - 639