Visual odometry combined with depth estimation network of improved dense block and multi-view geometry

被引:0
作者
Peng D.-G. [1 ,2 ]
Ouyang H.-L. [1 ]
Qi E.-J. [1 ,2 ]
Wang D.-H. [1 ]
机构
[1] College of Automation Engineering, Shanghai University of Electric Power, Shanghai
[2] Shanghai Engineering Research Center of Intelligent Management and Control for Power Process, Shanghai
来源
Kongzhi yu Juece/Control and Decision | 2023年 / 38卷 / 04期
关键词
dense block; depth estimation; multi-view geometry; optical flow estimation; unsupervised deep learning; visual odometry;
D O I
10.13195/j.kzyjc.2021.1264
中图分类号
学科分类号
摘要
An unsupervised monocular visual odometry based on the principle of multi-view geometry and effective combination of the convolutional neural network for image depth estimation and correspondences selection is proposed. Aiming at the problem that mainstream depth estimation networks tend to lose the shallow features of images, a depth estimation network based on improved dense blocks is constructed to effectively aggregate shallow features and improve the accuracy of image depth estimation. The odometry uses the depth estimation network to accurately predict the depth of the monocular image, uses the optical flow network to obtain forward-backward optical flow, and select high-quality correspondences based on the principle of forward and backward optical flow consistency. The initial pose and calculated depth are obtained by using multi-view geometric principles and optimization methods, and a 6-degree-of-freedom pose with the fixed global scale is obtained through a specific scale alignment principle. At the same time, in order to improve the network’s ability to learn scene details and the information of weak texture regions, the feature metric loss based on feature map synthesis is combined into the network loss function. On the KITTI Odometry dataset, the depth estimation under different thresholds has achieved accuracy rates of 85.9 %, 95.8 %, and 97.2 %, and the absolute trajectory error of the odometry evaluation on the 09 and 10 sequences is 0.007m. Experimental results show the effectiveness and accuracy of the proposed method, and prove that it is superior to the existing methods on the task of visual odometry. © 2023 Northeast University. All rights reserved.
引用
收藏
页码:980 / 988
页数:8
相关论文
共 24 条
  • [1] Zhou H Z, Ummenhofer B, Brox T., DeepTAM: Deep tracking and mapping with convolutional neural networks, International Journal of Computer Vision, 128, 3, pp. 756-769, (2020)
  • [2] Konda K, Memisevic R., Learning visual odometry with a convolutional network, Proceedings of the 10 th International Conference on Computer Vision Theory and Applications, pp. 486-490, (2015)
  • [3] Wang S, Clark R, Wen H K, Et al., DeepVO: Towards end-to-end visual odometry with deep recurrent convolutional neural networks, 2017 IEEE International Conference on Robotics and Automation, pp. 2043-2050, (2017)
  • [4] Zhou T H, Brown M, Snavely N, Et al., Unsupervised learning of depth and ego-motion from video, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), pp. 6612-6619, (2017)
  • [5] Bian J W, Li Z C, Wang N Y, Et al., Unsupervised scale-consistent depth and ego-motion learning from monocular video, (2019)
  • [6] Huang G, Liu Z, van der Maaten L, Et al., Densely connected convolutional networks, 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2261-2269, (2017)
  • [7] Guo Z B, Yang M K, Chen N H, Et al., LightVO: Lightweight inertial-assisted monocular visual odometry with dense neural networks, 2019 IEEE Global Communications Conference, pp. 1-6, (2019)
  • [8] Huang K, Qu X T, Chen S Q, Et al., Superb monocular depth estimation based on transfer learning and surface normal guidance, Sensors, 20, 17, pp. 4856-4878, (2020)
  • [9] Yu C, Liu Z X, Liu J X, Et al., Dense-loop: A loop closure detection method for visual SLAM using DenseNet features, The 1st International Workshop on the Semantic Descritor, Semantic Modelingard Mapping for Humanlike Perceptionand Navigation of Mobile Robots towards Large Scale Long-Term Autonomy (SDMM19), pp. 27-37, (2019)
  • [10] Fooladgar F, Kasaei S., Lightweight residual densely connected convolutional neural network, Multimedia Tools and Applications, 79, 35, pp. 25571-25588, (2020)