A visual questioning answering approach to enhance robot localization in indoor environments

被引:0
作者
Pena-Narvaez, Juan Diego [1 ]
Martin, Francisco [2 ]
Guerrero, Jose Miguel [2 ]
Perez-Rodriguez, Rodrigo [2 ]
机构
[1] Rey Juan Carlos Univ, Int Doctoral Sch, Signal Theory Commun Telemat Syst & Computat Dept, Intelligent Robot Lab, Fuenlabrada, Spain
[2] Rey Juan Carlos Univ, Intelligent Robot Lab, Signal Theory Commun Telemat Syst & Computat Dept, Fuenlabrada, Spain
关键词
visual question answering; robot localization; robot navigation; semantic map; robot mapping;
D O I
10.3389/fnbot.2023.1290584
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Navigating robots with precision in complex environments remains a significant challenge. In this article, we present an innovative approach to enhance robot localization in dynamic and intricate spaces like homes and offices. We leverage Visual Question Answering (VQA) techniques to integrate semantic insights into traditional mapping methods, formulating a novel position hypothesis generation to assist localization methods, while also addressing challenges related to mapping accuracy and localization reliability. Our methodology combines a probabilistic approach with the latest advances in Monte Carlo Localization methods and Visual Language models. The integration of our hypothesis generation mechanism results in more robust robot localization compared to existing approaches. Experimental validation demonstrates the effectiveness of our approach, surpassing state-of-the-art multi-hypothesis algorithms in both position estimation and particle quality. This highlights the potential for accurate self-localization, even in symmetric environments with large corridor spaces. Furthermore, our approach exhibits a high recovery rate from deliberate position alterations, showcasing its robustness. By merging visual sensing, semantic mapping, and advanced localization techniques, we open new horizons for robot navigation. Our work bridges the gap between visual perception, semantic understanding, and traditional mapping, enabling robots to interact with their environment through questions and enrich their map with valuable insights. The code for this project is available on GitHub https://github.com/juandpenan/topology_nav_ros2.
引用
收藏
页数:13
相关论文
共 32 条
  • [1] Ahn M, 2022, Arxiv, DOI arXiv:2204.01691
  • [2] OG-SGG: Ontology-Guided Scene Graph Generation-A Case Study in Transfer Learning for Telepresence Robotics
    Amodeo, Fernando
    Caballero, Fernando
    Diaz-Rodriguez, Natalia
    Merino, Luis
    [J]. IEEE ACCESS, 2022, 10 : 132564 - 132583
  • [3] Carion Nicolas, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12346), P213, DOI 10.1007/978-3-030-58452-8_13
  • [4] Cassandra AR, 1996, IROS 96 - PROCEEDINGS OF THE 1996 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS - ROBOTIC INTELLIGENCE INTERACTING WITH DYNAMIC WORLDS, VOLS 1-3, P963, DOI 10.1109/IROS.1996.571080
  • [5] Robust 2D Indoor Localization through Laser SLAM and Visual SLAM Fusion
    Chan, Shao-Hung
    Wu, Ping-Tsang
    Fu, Li-Chen
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 1263 - 1268
  • [6] A Review of Visual-LiDAR Fusion based Simultaneous Localization and Mapping
    Debeunne, Cesar
    Vivet, Damien
    [J]. SENSORS, 2020, 20 (07)
  • [7] Deng YH, 2021, ROBOT SCI SYS
  • [8] Driess D, 2023, Arxiv, DOI [arXiv:2303.03378, 10.48550/arXiv.2303.03378]
  • [9] ENGELSON SP, 1992, 1992 IEEE INTERNATIONAL CONF ON ROBOTICS AND AUTOMATION : PROCEEDINGS, VOLS 1-3, P2555, DOI 10.1109/ROBOT.1992.220057
  • [10] Portable Multi-Hypothesis Monte Carlo Localization for Mobile Robots
    Garcia, Alberto
    Martin, Francisco
    Guerrero, Jose Miguel
    Rodriguez, Francisco J.
    Matellan, Vicente
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 1933 - 1939