On Embodied Visual Navigation in Real Environments Through Habitat

被引:10
作者
Rosano, Marco [1 ,3 ]
Furnari, Antonino [1 ]
Gulino, Luigi [3 ]
Farinella, Giovanni Maria [1 ,2 ]
机构
[1] Univ Catania, Dept Math & Comp Sci, FPV, IPLAB, Catania, Italy
[2] CNR, Cognit Robot & Social Sensing Lab, ICAR, Palermo, Italy
[3] OrangeDev Srl, Florence, Italy
来源
2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) | 2021年
关键词
D O I
10.1109/ICPR48806.2021.9413026
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual navigation models based on deep learning can learn effective policies when trained on large amounts of visual observations through reinforcement learning. Unfortunately, collecting the required experience in the real world requires the deployment of a robotic platform, which is expensive and time-consuming. To deal with this limitation, several simulation platforms have been proposed in order to train visual navigation policies on virtual environments efficiently. Despite the advantages they offer, simulators present a limited realism in terms of appearance and physical dynamics, leading to navigation policies that do not generalize in the real world. In this paper, we propose a tool based on the Habitat simulator which exploits real world images of the environment, together with sensor and actuator noise models, to produce more realistic navigation episodes. We perform a range of experiments to assess the ability of such policies to generalize using virtual and real-world images, as well as observations transformed with unsupervised domain adaptation approaches. We also assess the impact of sensor and actuation noise on the navigation performance and investigate whether it allows to learn more robust navigation policies. We show that our tool can effectively help to train and evaluate navigation policies on real-world observations without running navigation episodes in the real world.
引用
收藏
页码:9740 / 9747
页数:8
相关论文
共 28 条
[1]  
Anderson Peter, 2018, CoRR
[2]  
[Anonymous], 2017, IEEE I CONF COMP VIS, DOI DOI 10.1109/ICCV.2017.244
[3]   Visual navigation for mobile robots: A survey [J].
Bonin-Font, Francisco ;
Ortiz, Alberto ;
Oliver, Gabriel .
JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2008, 53 (03) :263-296
[4]   Matterport3D: Learning from RGB-D Data in Indoor Environments [J].
Chang, Angel ;
Dai, Angela ;
Funkhouser, Thomas ;
Halber, Maciej ;
Niessner, Matthias ;
Savva, Manolis ;
Song, Shuran ;
Zeng, Andy ;
Zhang, Yinda .
PROCEEDINGS 2017 INTERNATIONAL CONFERENCE ON 3D VISION (3DV), 2017, :667-676
[5]  
Chaplot Devendra Singh, 2020, INT C LEARN REPR ICL
[6]  
Chen T., 2019, PROC INT C LEARN REP
[7]   RoboTHOR: An Open Simulation-to-Real Embodied AI Platform [J].
Deitke, Matt ;
Han, Winson ;
Herrasti, Alvaro ;
Kembhavi, Aniruddha ;
Kolve, Eric ;
Mottaghi, Roozbeh ;
Salvador, Jordi ;
Schwenk, Dustin ;
VanderBilt, Eli ;
Wallingford, Matthew ;
Weihs, Luca ;
Yatskar, Mark ;
Farhadi, Ali .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3161-3171
[8]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[9]   Cognitive Mapping and Planning for Visual Navigation [J].
Gupta, Saurabh ;
Davidson, James ;
Levine, Sergey ;
Sukthankar, Rahul ;
Malik, Jitendra .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7272-7281
[10]  
Hoffman J, 2018, PR MACH LEARN RES, V80