Context-Enhanced Meta-Reinforcement Learning with Data-Reused Adaptation for Urban Autonomous Driving

被引:2
作者
Deng, Qi [1 ]
Zhao, Yaqian [1 ]
Li, Rengang [1 ]
Hu, Qifu [1 ]
Liu, Tiejun [1 ]
Li, Ruyang [1 ]
机构
[1] Inspur Elect Informat Ind Co Ltd, Beijing, Peoples R China
来源
2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN | 2023年
关键词
Autonomous driving; Meta-reinforcement learning; State encoding; Policy adaptation;
D O I
10.1109/IJCNN54540.2023.10191187
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autonomous driving (AD) has experienced rapid development in recent years, and the reinforcement learning (RL) pipeline in trial-and-error manner can surpass human driving ability. However, the poor performance in sample efficiency and generalization limits RL applying in the challenging urban traffic scenarios. In this paper, we build a context-enhanced meta-RL framework with data-reused adaptation for challenging urban AD. At both the meta-learning and adaptation stages, the context-enhanced state representation is designed to reduce the perceptual gap in variant urban scenarios, improving the sample efficiency and robustness. At adaptation stage, the metatraining data with context-enhanced features are reused through propensity estimation to constrain the optimization objective of new tasks, aiming to maintain the good driving performance of meta-trained policy and fast adapt to the new tasks. Extensive experiments are conducted in CARLA simulator with various urban environments and task settings. The learning curves and quantitative comparisons validate the good sample efficiency and generalization of our proposed method, with state-of-the-art driving performance on urban AD benchmarks.
引用
收藏
页数:8
相关论文
共 36 条
[1]  
[Anonymous], 2017, 31 INT C NEUR INF PR, DOI DOI 10.5555/3294771.3294875
[2]   Reinforcement Learning, Fast and Slow [J].
Botvinick, Matthew ;
Ritter, Sam ;
Wang, Jane X. ;
Kurth-Nelson, Zeb ;
Blundell, Charles ;
Hassabis, Demis .
TRENDS IN COGNITIVE SCIENCES, 2019, 23 (05) :408-422
[3]   Learning to drive from a world on rails [J].
Chen, Dian ;
Koltun, Vladlen ;
Krahenbuhl, Philipp .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :15570-15579
[4]   Learning from Simulation, Racing in Reality [J].
Chisari, Eugenio ;
Liniger, Alexander ;
Rupenyan, Alisa ;
Van Gool, Luc ;
Lygeros, John .
2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, :8046-8052
[5]   NEAT: Neural Attention Fields for End-to-End Autonomous Driving [J].
Chitta, Kashyap ;
Prakash, Aditya ;
Geiger, Andreas .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :15773-15783
[6]  
Dosovitskiy A, 2017, PR MACH LEARN RES, V78
[7]  
Duan Y., 2016, Rl2: Fast reinforcement learning via slow reinforcement learning
[8]  
Finn C, 2017, PR MACH LEARN RES, V70
[9]  
Finn Chelsea, 2017, P MACHINE LEARNING R, V78
[10]  
Fujimoto S, 2019, PR MACH LEARN RES, V97