Deep Learning-Based Visual Navigation Algorithms for Mobile Robots: A Comprehensive Study

被引:0
作者
Yu W. [1 ]
Tian X. [1 ]
机构
[1] Xi'an Siyuan University, Shaanxi, Xi'an
关键词
deep learning; mobile robots; Neo-model; ResNet; split attention; visual navigation;
D O I
10.20532/cit.2022.1005689
中图分类号
学科分类号
摘要
This research addresses the challenges faced by mobile robots in efficiently navigating complex environments. A novel approach is proposed, leveraging deep learning techniques, and introducing the Neo model. The method combines Split Attention with the ResNeSt50 network to enhance the recognition accuracy of key features in the observed images. Furthermore, improvements have been made in the loss calculation method to improve navigation accuracy across different scenarios. Evaluations conducted on AI2THOR and active vision datasets demonstrate that the improved model achieves higher average navigation accuracy (92.3%) in scene 4 compared to other methods. The success rate of navigation reached 36.8%, accompanied by a 50% reduction in ballistic length. Additionally, compared to HAUSR and LSTM-Nav, this technology significantly reduced collision rates to 0.01 and reduced time consumption by over 8 seconds. The research methodology addresses navigation model accuracy, speed, and generalization issues, thus making significant advancements for intelligent autonomous robots. ACM CCS (2012) Classification: Computing methodologies → Machine learning → Machine learning algorithms Artificial intelligence → Computer vision → Vision for robotics. © 2022, University of Zagreb Faculty of Electrical Engineering and Computing. All rights reserved.
引用
收藏
页码:257 / 273
页数:16
相关论文
共 29 条
[1]  
Kulhanek J., Et al., Visual Navigation in Real-World Indoor Envi-ronments Using End-toEnd Deep Reinforcement Learning, IEEE Robotics and Auto-mation Letters, 6, 3, pp. 4345-4352, (2021)
[2]  
Morad S. D., Et al., Embodied Visual Navigation With Automatic Curriculum Learning in Real Environments, IEEE Robotics and Automation Letters, 6, 2, pp. 683-690, (2021)
[3]  
Fang Y., Et al., ST-SIGMA: Spatio-temporal Semantics and Interaction Graph Aggregation for Multi-agent Perception and Trajectory Forecasting, CAAI Transactions on Intelligence Technology, 7, 4, pp. 744-757, (2022)
[4]  
Wu Q., Et al., Reinforcement Learning-Based Visual Navigation With Information-Theoretic Regularization, IEEE Robotics and Automation Letters, 6, 2, pp. 731-738, (2021)
[5]  
Lyu Y., Et al., Improving Target-driven Visual Navigation with Attention on 3D Spatial Relationships, Neural Processing Letters, 54, 5, pp. 3979-3998, (2022)
[6]  
Tolani V., Et al., Visual Navigation Among Humans With Optimal Control as a Supervisor, IEEE Robotics and Automation Letters, 6, 2, pp. 2288-2295, (2021)
[7]  
Zhao Y., Et al., Robot Visual Navigation Estimation and Target Localization Based on Neural Network, Paladyn, Journal of Behavioral Robotics, 13, 1, pp. 76-83, (2022)
[8]  
Fang Q., Et al., Target-driven Visual Navigation in Indoor Scenes using Reinforcement Learning and Imitation Learning, CAAI Transactions on Intelligence Technology, 7, 2, pp. 167-176, (2022)
[9]  
Guan T., Et al., GA-Nav: Efficient Terrain Segmentation for Robot Navigation in Unstructured Outdoor Environments, IEEE Robotics and Automation Letters, 7, 3, pp. 8138-8145, (2022)
[10]  
Sang H., Et al., A Novel Neural Multi-Store Memory Network for Autonomous Visual Navigation in Unknown Environment, IEEE Robotics and Automation Letters, 7, 2, pp. 2039-2046, (2022)