Scale-Aware Visual-Inertial Depth Estimation and Odometry Using Monocular Self-Supervised Learning

被引:4
|
作者
Lee, Chungkeun [1 ]
Kim, Changhyeon [2 ]
Kim, Pyojin [3 ]
Lee, Hyeonbeom [4 ]
Kim, H. Jin [5 ]
机构
[1] Seoul Natl Univ, Inst Adv Aerosp Technol, Seoul 08826, South Korea
[2] Seoul Natl Univ, Automation & Syst Res Inst, Seoul 08826, South Korea
[3] Sookmyung Womens Univ, Dept Mech Syst Engn, Seoul 04312, South Korea
[4] Kyungpook Natl Univ, Sch Elect & Elect Engn, Daegu 37224, South Korea
[5] Seoul Natl Univ, Dept Mech & Aerosp Engn, Seoul 08826, South Korea
基金
新加坡国家研究基金会;
关键词
Odometry; Deep learning; Loss measurement; Depth measurement; Cameras; Self-supervised learning; Coordinate measuring machines; monocular depth estimation; self-supervised learning; visual-inertial odometry;
D O I
10.1109/ACCESS.2023.3252884
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
For real-world applications with a single monocular camera, scale ambiguity is an important issue. Because self-supervised data-driven approaches that do not require additional data containing scale information cannot avoid the scale ambiguity, state-of-the-art deep-learning-based methods address this issue by learning the scale information from additional sensor measurements. In that regard, inertial measurement unit (IMU) is a popular sensor for various mobile platforms due to its lightweight and inexpensiveness. However, unlike supervised learning that can learn the scale from the ground-truth information, learning the scale from IMU is challenging in a self-supervised setting. We propose a scale-aware monocular visual-inertial depth estimation and odometry method with end-to-end training. To learn the scale from the IMU measurements with end-to-end training in the monocular self-supervised setup, we propose a new loss function named as preintegration loss function, which trains scale-aware ego-motion by comparing the ego-motion integrated from IMU measurement and predicted ego-motion. Since the gravity and the bias should be compensated to obtain the ego-motion by integrating IMU measurements, we design a network to predict the gravity and the bias in addition to the ego-motion and the depth map. The overall performance of the proposed method is compared to state-of-the-art methods in the popular outdoor driving dataset, i.e., KITTI dataset, and the author-collected indoor driving dataset. In the KITTI dataset, the proposed method shows competitive performance compared with state-of-the-art monocular depth estimation and odometry methods, i.e., root-mean-square error of 5.435 m in the KITTI Eigen split and absolute trajectory error of 22.46 m and 0.2975 degrees in the KITTI odometry 09 sequence. Different from other up-to-scale monocular methods, the proposed method can estimate the metric-scaled depth and camera poses. Additional experiments on the author-collected indoor driving dataset qualitatively confirm the accurate performance of metric-depth and metric pose estimations.
引用
收藏
页码:24087 / 24102
页数:16
相关论文
共 50 条
  • [31] Revisiting Self-supervised Monocular Depth Estimation
    Kim, Ue-Hwan
    Lee, Gyeong-Min
    Kim, Jong-Hwan
    ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS 6, 2022, 429 : 336 - 350
  • [32] Online Self-Supervised Monocular Visual Odometry for Ground Vehicles
    Lee, Rhoram
    Daniilidis, Kostas
    Lee, Daniel D.
    2015 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2015, : 5232 - 5238
  • [33] Keypoint Heatmap Guided Self-Supervised Monocular Visual Odometry
    Haixin Xiu
    Yiyou Liang
    Hui Zeng
    Journal of Intelligent & Robotic Systems, 2022, 105
  • [34] Keypoint Heatmap Guided Self-Supervised Monocular Visual Odometry
    Xiu, Haixin
    Liang, Yiyou
    Zeng, Hui
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2022, 105 (04)
  • [35] MotionHint: Self-Supervised Monocular Visual Odometry with Motion Constraints
    Wang, Cong
    Wang, Yu-Ping
    Manocha, Dinesh
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022,
  • [36] Learning Occlusion-aware Coarse-to-Fine Depth Map for Self-supervised Monocular Depth Estimation
    Zhou, Zhengming
    Dong, Qiulei
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 6386 - 6395
  • [37] Multimodal Scale Consistency and Awareness for Monocular Self-Supervised Depth Estimation
    Chawla, Hemang
    Varma, Arnav
    Arani, Elahe
    Zonooz, Bahram
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 5140 - 5146
  • [38] Spatial-Aware Dynamic Lightweight Self-Supervised Monocular Depth Estimation
    Song, Linna
    Shi, Dianxi
    Xia, Jianqiang
    Ouyang, Qianying
    Qiao, Ziteng
    Jin, Songchang
    Yang, Shaowu
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (01): : 883 - 890
  • [39] Absolute Monocular Depth Estimation on Robotic Visual and Kinematics Data via Self-Supervised Learning
    Wei, Ruofeng
    Li, Bin
    Zhong, Fangxun
    Mo, Hangjie
    Dou, Qi
    Liu, Yun-Hui
    Sun, Dong
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 1 - 14
  • [40] Visual-Inertial Odometry with Robust Initialization and Online Scale Estimation
    Hong, Euntae
    Lim, Jongwoo
    SENSORS, 2018, 18 (12)