Learning Model Predictive Controllers with Real-Time Attention for Real-World Navigation

被引:0
作者
Xiao, Xuesu [1 ,2 ]
Zhang, Tingnan [3 ]
Choromanski, Krzysztof [3 ]
Lee, Edward [3 ]
Francis, Anthony [3 ]
Varley, Jake [3 ]
Tu, Stephen [3 ]
Singh, Sumeet [3 ]
Xu, Peng [3 ]
Xia, Fei [3 ]
Persson, Sven Mikael [2 ]
Kalashnikov, Dmitry [3 ]
Takayama, Leila [4 ]
Frostig, Roy [3 ]
Tan, Jie [3 ]
Parada, Carolina [3 ]
Sindhwani, Vikas [3 ]
机构
[1] George Mason Univ, Fairfax, VA 22030 USA
[2] Everyday Robots, Mountain View, CA 94043 USA
[3] Robot Google, Mountain View, CA USA
[4] Hoku Labs, Santa Cruz, CA USA
来源
CONFERENCE ON ROBOT LEARNING, VOL 205 | 2022年 / 205卷
关键词
Model Predictive Control; Transformers; Performers; Highly-Constrained Navigation; Social Navigation; Learning-based Control; OFF-ROAD NAVIGATION; END-TO-END; NEWTON METHOD;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite decades of research, existing navigation systems still face real-world challenges when deployed in the wild, e.g., in cluttered home environments or in human-occupied public spaces. To address this, we present a new class of implicit control policies combining the benefits of imitation learning with the robust handling of system constraints from Model Predictive Control (MPC). Our approach, called Performer-MPC,(1) uses a learned cost function parameterized by vision context embeddings provided by Performers-a low-rank implicit-attention Transformer. We jointly train the cost function and construct the controller relying on it, effectively solving end-to-end the corresponding bi-level optimization problem. We show that the resulting policy improves standard MPC performance by leveraging a few expert demonstrations of the desired navigation behavior in different challenging real-world scenarios. Compared with a standard MPC policy, Performer-MPC achieves >40% better goal reached in cluttered environments and >65% better on social metrics when navigating around humans.
引用
收藏
页码:1708 / 1721
页数:14
相关论文
共 98 条
[1]  
Agrawal A, 2019, Arxiv, DOI arXiv:1912.09529
[2]  
Agrawal A, 2019, ADV NEUR IN, V32
[3]  
Ahn M, 2022, Arxiv, DOI [arXiv:2204.01691, DOI 10.48550/ARXIV.2204.01691]
[4]  
Alayrac JB, 2022, ADV NEUR IN
[5]  
Amazon, 2022, Meet scout
[6]  
Amos Brandon, 2018, Advances in Neural Information Processing Systems, V31
[7]   Learning for Autonomous Navigation Advances in Machine Learning for Rough Terrain Mobility [J].
Bagnell, James Andrew ;
Bradley, David ;
Silver, David ;
Sofman, Boris ;
Stentz, Anthony .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 2010, 17 (02) :74-84
[8]   Autonomous Off-Road Navigation with End-to-End Learning for the LAGR Program [J].
Bajracharya, Max ;
Howard, Andrew ;
Matthies, Larry H. ;
Tang, Benyang ;
Turmon, Michael .
JOURNAL OF FIELD ROBOTICS, 2009, 26 (01) :3-25
[9]  
Bertsekas D.P., 1995, NONLINEAR PROGRAMMIN, DOI [10.1057/palgrave.jors.2600425, DOI 10.1057/PALGRAVE.JORS.2600425]
[10]  
Bhardwaj Mohak, 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA), P10598, DOI 10.1109/ICRA40945.2020.9197260