RGB-D-E: Event Camera Calibration for Fast 6-DOF Object Tracking

被引:15
作者
Dubeau, Etienne [1 ]
Garon, Mathieu [1 ]
Debaque, Benoit [2 ]
de Charette, Raoul [3 ]
Lalonde, Jean-Francois [1 ]
机构
[1] Univ Laval, Quebec City, PQ, Canada
[2] Thales Digital Solut, Ottawa, ON, Canada
[3] INRIA, Rocquencourt, France
来源
2020 IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY (ISMAR 2020) | 2020年
基金
加拿大自然科学与工程研究理事会;
关键词
Event camera; Calibration; 6-DOF Object tracking; Augmented reality;
D O I
10.1109/ISMAR50242.2020.00034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Augmented reality devices require multiple sensors to perform various tasks such as localization and tracking. Currently, popular cameras are mostly frame-based (e.g. RGB and Depth) which impose a high data bandwidth and power usage. With the necessity for low power and more responsive augmented reality systems, using solely frame-based sensors imposes limits to the various algorithms that needs high frequency data from the environement. As such, event-based sensors have become increasingly popular due to their low power, bandwidth and latency, as well as their very high frequency data acquisition capabilities. In this paper, we propose, for the first time, to use an event-based camera to increase the speed of 3D object tracking in 6 degrees of freedom. This application requires handling very high object speed to convey compelling AR experiences. To this end, we propose a new system which combines a recent RGB-D sensor (Kinect Azure) with an event camera (DAVIS346). We develop a deep learning approach, which combines an existing RGB-D network along with a novel event-based network in a cascade fashion, and demonstrate that our approach significantly improves the robustness of a state-of-the-art frame-based 6-DOF object tracker using our RGB-D-E pipeline. Our code and our RGB-D-E evaluation dataset are available at https://github.com/lvsn/rgbde_tracking.
引用
收藏
页码:127 / 135
页数:9
相关论文
共 45 条
[1]   Real-Time 6DOF Pose Relocalization for Event Cameras with Stacked Spatial LSTM Networks [J].
Anh Nguyen ;
Thanh-Toan Do ;
Caldwell, Darwin G. ;
Tsagarakis, Nikos G. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, :1638-1645
[2]  
[Anonymous], 2016, ARXIV
[3]  
Bryner S, 2019, IEEE INT CONF ROBOT, P325, DOI [10.1109/ICRA.2019.8794255, 10.1109/icra.2019.8794255]
[4]   Calibrate Multiple Consumer RGB-D Cameras for Low-Cost and Efficient 3D Indoor Mapping [J].
Chen, Chi ;
Yang, Bisheng ;
Song, Shuang ;
Tian, Mao ;
Li, Jianping ;
Dai, Wenxia ;
Fang, Lina .
REMOTE SENSING, 2018, 10 (02)
[5]  
Delmerico J, 2019, IEEE INT CONF ROBOT, P6713, DOI [10.1109/ICRA.2019.8793887, 10.1109/icra.2019.8793887]
[6]   RANDOM SAMPLE CONSENSUS - A PARADIGM FOR MODEL-FITTING WITH APPLICATIONS TO IMAGE-ANALYSIS AND AUTOMATED CARTOGRAPHY [J].
FISCHLER, MA ;
BOLLES, RC .
COMMUNICATIONS OF THE ACM, 1981, 24 (06) :381-395
[7]   Event-Based, 6-DOF Camera Tracking from Photometric Depth Maps [J].
Gallego, Guillermo ;
Lund, Jon E. A. ;
Mueggler, Elias ;
Rebecq, Henri ;
Delbruck, Tobi ;
Scaramuzza, Davide .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (10) :2402-2412
[8]   A Framework for Evaluating 6-DOF Object Trackers [J].
Garon, Mathieu ;
Laurendeau, Denis ;
Lalonde, Jean-Francois .
COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 :608-623
[9]   Deep 6-DOF Tracking [J].
Garon, Mathieu ;
Lalonde, Jean-Francois .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2017, 23 (11) :2410-2418
[10]   End-to-End Learning of Representations for Asynchronous Event-Based Data [J].
Gehrig, Daniel ;
Loquercio, Antonio ;
Derpanis, Konstantinos G. ;
Scaramuzza, Davide .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :5632-5642