Geometry-Aware Network for Unsupervised Learning of Monocular Camera's Ego-Motion

被引:2
|
作者
Zhou, Beibei [1 ,2 ]
Xie, Jin [1 ,2 ]
Jin, Zhong [1 ,2 ]
Kong, Hui [3 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, PCA Lab, Key Lab Intelligent Percept & Syst High Dimens Inf, Nanjing 210094, Jiangsu, Peoples R China
[2] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Jiangsu Key Laboratoryof Image & Video Understandi, Nanjing 210094, Jiangsu, Peoples R China
[3] Univ Macau, Dept Electromech Engn EME, State Key Lab Internet Things Smart City SKL IOTSC, Macau, Peoples R China
关键词
Index Terms-Monocular visual odometry; geometry-aware; point clouds; visual appearance; 6-DoF poses; VISUAL ODOMETRY; DEPTH;
D O I
10.1109/TITS.2023.3298715
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Deep neural networks have been shown to be effective for unsupervised monocular visual odometry that can predict the camera's ego-motion based on an input of monocular video sequence. However, most existing unsupervised monocular methods haven't fully exploited the extracted information from both local geometric structure and visual appearance of the scenes, resulting in degraded performance. In this paper, a novel geometry-aware network is proposed to predict the camera's ego-motion by learning representations in both 2D and 3D space. First, to extract geometry-aware features, we design an RGB-PointCloud feature fusion module to capture information from both geometric structure and the visual appearance of the scenes by fusing local geometric features from depth-map-derived point clouds and visual features from RGB images. Furthermore, the fusion module can adaptively allocate different weights to the two types of features to emphasize important regions. Then, we devise a relevant feature filtering module to build consistency between the two views and preserve informative features with high relevance. It can capture the correlation of frame pairs in the feature-embedding space by attention mechanisms. Finally, the obtained features are fed into the pose estimator to recover the 6-DoF poses of the camera. Extensive experiments show that our method achieves promising results among the unsupervised monocular deep learning methods on the KITTI odometry and TUM-RGBD datasets.
引用
收藏
页码:14226 / 14236
页数:11
相关论文
共 50 条
  • [31] Unsupervised Learning of Depth and Ego-Motion from Monocular Video Using 3D Geometric Constraints
    Mahjourian, Reza
    Wicke, Martin
    Angelova, Anelia
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5667 - 5675
  • [32] Epipolar Geometry based Learning of Multi-view Depth and Ego-Motion from Monocular Sequences
    Prasad, Vignesh
    Das, Dipanjan
    Bhowmick, Brojeshwar
    ELEVENTH INDIAN CONFERENCE ON COMPUTER VISION, GRAPHICS AND IMAGE PROCESSING (ICVGIP 2018), 2018,
  • [33] GOPE: Geometry-Aware Optimal Viewpoint Path Estimation Using a Monocular Camera
    Kim, Nuri
    Choi, Yunho
    Kang, Minjae
    Oh, Songhwai
    2020 20TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS), 2020, : 1062 - 1067
  • [34] ToF Camera Ego-Motion Estimation
    Ratshidaho, Terence
    Tapamo, Jules Raymond
    Claassens, Jonathan
    Govender, Natasha
    2012 5TH ROBOTICS AND MECHATRONICS CONFERENCE OF SOUTH AFRICA (ROBOMECH), 2012,
  • [35] Unsupervised Learning of Efficient Geometry-Aware Neural Articulated Representations
    Noguchi, Atsuhiro
    Sun, Xiao
    Lin, Stephen
    Harada, Tatsuya
    COMPUTER VISION - ECCV 2022, PT XVII, 2022, 13677 : 597 - 614
  • [36] Unsupervised Learning of Depth and Ego-Motion from Cylindrical Panoramic Video
    Sharma, Alisha
    Ventura, Jonathan
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND VIRTUAL REALITY (AIVR), 2019, : 58 - 65
  • [37] Influence of Neural Network Receptive Field on Monocular Depth and Ego-Motion Estimation
    Linok, S. A.
    Yudin, D. A.
    OPTICAL MEMORY AND NEURAL NETWORKS, 2023, 32 (Suppl 2) : S206 - S213
  • [38] Influence of Neural Network Receptive Field on Monocular Depth and Ego-Motion Estimation
    S. A. Linok
    D. A. Yudin
    Optical Memory and Neural Networks, 2023, 32 : S206 - S213
  • [39] Robust estimation of camera ego-motion parameters
    Wang, Jian-Ming
    Yan, Zhi-Jie
    Duan, Xiao-Jie
    Dou, Ru-Zhen
    Leng, Yu
    Hongwai yu Jiguang Gongcheng/Infrared and Laser Engineering, 2010, 39 (06): : 1168 - 1172
  • [40] A new Method on Camera Ego-motion Estimation
    Yuan, Ding
    Yu, Yalong
    2013 6TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING (CISP), VOLS 1-3, 2013, : 651 - 656