Where to Look Next: Learning Viewpoint Recommendations for Informative Trajectory Planning

被引:18
作者
Lodel, Max [1 ]
Brito, Bruno [1 ]
Serra-Gomez, Alvaro [1 ]
Ferranti, Laura [1 ]
Babuska, Robert [2 ]
Alonso-Mora, Javier [1 ]
机构
[1] Delft Univ Technol, Dept Cognit Robot CoR, NL-2628 CD Delft, Netherlands
[2] Czech Tech Univ, CIIRC, Prague, Czech Republic
来源
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022) | 2022年
关键词
NAVIGATION; EXPLORATION;
D O I
10.1109/ICRA46639.2022.9812190
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Search missions require motion planning and navigation methods for information gathering that continuously replan based on new observations of the robot's surroundings. Current methods for information gathering, such as Monte Carlo Tree Search, are capable of reasoning over long horizons, but they are computationally expensive. An alternative for fast online execution is to train, offline, an information gathering policy, which indirectly reasons about the information value of new observations. However, these policies lack safety guarantees and do not account for the robot dynamics. To overcome these limitations we train an information-aware policy via deep reinforcement learning, that guides a receding-horizon trajectory optimization planner. In particular, the policy continuously recommends a reference viewpoint to the local planner, such that the resulting dynamically feasible and collision-free trajectories lead to observations that maximize the information gain and reduce the uncertainty about the environment. In simulation tests in previously unseen environments, our method consistently outperforms greedy next-best-view policies and achieves competitive performance compared to Monte Carlo Tree Search, in terms of information gains and coverage time, with a reduction in execution time by three orders of magnitude.
引用
收藏
页码:4466 / 4472
页数:7
相关论文
共 40 条
[1]  
Atanasov N, 2014, IEEE INT CONF ROBOT, P6447, DOI 10.1109/ICRA.2014.6907811
[2]  
Bengio Y., 2009, ACM INT C P SER, V382
[3]   Dec-MCTS: Decentralized planning for multi-robot active perception [J].
Best, Graeme ;
Cliff, Oliver M. ;
Patten, Timothy ;
Mettu, Ramgopal R. ;
Fitch, Robert .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2019, 38 (2-3) :316-337
[4]   Receding horizon path planning for 3D exploration and surface inspection [J].
Bircher, Andreas ;
Kamel, Mina ;
Alexis, Kostas ;
Oleynikova, Helen ;
Siegwart, Roland .
AUTONOMOUS ROBOTS, 2018, 42 (02) :291-306
[5]  
Bourgault F, 2002, 2002 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-3, PROCEEDINGS, P540, DOI 10.1109/IRDS.2002.1041446
[6]   Where to go Next: Learning a Subgoal Recommendation Policy for Navigation in Dynamic Environments [J].
Brito, Bruno ;
Everett, Michael ;
How, Jonathan P. ;
Alonso-Mora, Javier .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) :4616-4623
[7]   Model Predictive Contouring Control for Collision Avoidance in Unstructured Dynamic Environments [J].
Brito, Bruno ;
Floor, Boaz ;
Ferranti, Laura ;
Alonso-Mora, Javier .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (04) :4459-4466
[8]  
Cao C, 2021, ROBOT SCI SYS
[9]  
Chaplot Devendra Singh, 2020, INT C LEARN REPR
[10]  
Charrow B., 2015, Robotics: Science and Systems, V11, P3