Geometry-Aware Network for Unsupervised Learning of Monocular Camera's Ego-Motion

被引:2
|
作者
Zhou, Beibei [1 ,2 ]
Xie, Jin [1 ,2 ]
Jin, Zhong [1 ,2 ]
Kong, Hui [3 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, PCA Lab, Key Lab Intelligent Percept & Syst High Dimens Inf, Nanjing 210094, Jiangsu, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Jiangsu Key Laboratoryof Image & Video Understandi, Nanjing 210094, Jiangsu, Peoples R China
[3] Univ Macau, Dept Electromech Engn EME, State Key Lab Internet Things Smart City SKL IOTSC, Macau, Peoples R China
关键词
Index Terms-Monocular visual odometry; geometry-aware; point clouds; visual appearance; 6-DoF poses; VISUAL ODOMETRY; DEPTH;
D O I
10.1109/TITS.2023.3298715
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep neural networks have been shown to be effective for unsupervised monocular visual odometry that can predict the camera's ego-motion based on an input of monocular video sequence. However, most existing unsupervised monocular methods haven't fully exploited the extracted information from both local geometric structure and visual appearance of the scenes, resulting in degraded performance. In this paper, a novel geometry-aware network is proposed to predict the camera's ego-motion by learning representations in both 2D and 3D space. First, to extract geometry-aware features, we design an RGB-PointCloud feature fusion module to capture information from both geometric structure and the visual appearance of the scenes by fusing local geometric features from depth-map-derived point clouds and visual features from RGB images. Furthermore, the fusion module can adaptively allocate different weights to the two types of features to emphasize important regions. Then, we devise a relevant feature filtering module to build consistency between the two views and preserve informative features with high relevance. It can capture the correlation of frame pairs in the feature-embedding space by attention mechanisms. Finally, the obtained features are fed into the pose estimator to recover the 6-DoF poses of the camera. Extensive experiments show that our method achieves promising results among the unsupervised monocular deep learning methods on the KITTI odometry and TUM-RGBD datasets.
引用
收藏
页码:14226 / 14236
页数:11
相关论文
共 50 条
  • [1] Improving Unsupervised Learning of Monocular Depth and Ego-Motion via Stereo Network
    He, Mu
    Xie, Jin
    Yang, Jian
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2021, PT II, 2021, 13020 : 421 - 433
  • [2] Unsupervised monocular depth and ego-motion learning with structure and semantics
    Casser, Vincent
    Pirk, Soeren
    Mahjourian, Reza
    Angelova, Anelia
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 381 - 388
  • [3] Unsupervised Learning of Monocular Depth and Ego-Motion Using Multiple Masks
    Wang, Guangming
    Wang, Hesheng
    Liu, Yiling
    Chen, Weidong
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 4724 - 4730
  • [4] Unsupervised Learning of Depth and Ego-Motion from Continuous Monocular Images
    Wang, Zhuo
    Huang, Min
    Huang, Xiao-Long
    Ma, Fei
    Dou, Jia-Ming
    Lyu, Jian-Li
    Journal of Computers (Taiwan), 2021, 32 (06) : 38 - 51
  • [5] Unsupervised Learning of Monocular Depth and Ego-Motion using Conditional PatchGANs
    Vankadari, Madhu
    Kumar, Swagat
    Majumder, Anima
    Das, Kaushik
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5677 - 5684
  • [6] Unsupervised Learning of Monocular Depth and Ego-Motion in Outdoor/Indoor Environments
    Gao, Ruipeng
    Xiao, Xuan
    Xing, Weiwei
    Li, Chi
    Liu, Lei
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (17) : 16247 - 16258
  • [7] Monocular Ego-motion Estimation with a Compact Omnidirectional Camera
    Stuerzl, Wolfgang
    Burschka, Darius
    Suppa, Michael
    IEEE/RSJ 2010 INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2010), 2010, : 822 - 828
  • [8] Depth Estimation with Ego-Motion Assisted Monocular Camera
    Mansour M.
    Davidson P.
    Stepanov O.
    Raunio J.-P.
    Aref M.M.
    Piché R.
    Gyroscopy Navig., 3 (111-123): : 111 - 123
  • [9] Unsupervised Ego-Motion and Dense Depth Estimation with Monocular Video
    Xu, Yufan
    Wang, Yan
    Guo, Lei
    2018 IEEE 18TH INTERNATIONAL CONFERENCE ON COMMUNICATION TECHNOLOGY (ICCT), 2018, : 1306 - 1310
  • [10] Unsupervised Scale-consistent Depth and Ego-motion Learning from Monocular Video
    Bian, Jia-Wang
    Li, Zhichao
    Wang, Naiyan
    Zhan, Huangying
    Shen, Chunhua
    Cheng, Ming-Ming
    Reid, Ian
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32