Residual Policy Learning Facilitates Efficient Model-Free Autonomous Racing

被引:25
作者
Zhang, Ruiqi [1 ]
Hou, Jing [1 ]
Chen, Guang [1 ,2 ]
Li, Zhijun [3 ]
Chen, Jianxiao [1 ]
Knoll, Alois [2 ]
机构
[1] Tongji Univ, Sch Automot Studies, Shanghai 201804, Peoples R China
[2] Tech Univ Munich, Dept Informat, Munich, Germany
[3] Univ Sci & Technol China, Wearable Robot & Autonomous Syst Lab, Hefei 230022, Peoples R China
基金
欧盟地平线“2020”; 中国国家自然科学基金;
关键词
Autonomous vehicle navigation; motion and path planning; reinforcement learning; PREDICTIVE CONTROL; AVOIDANCE;
D O I
10.1109/LRA.2022.3192770
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Motion planning for autonomous racing is a challenging task due to the safety requirement while driving aggressively. Most previous solutions utilize the prior information or depend on complex dynamics modeling. Classical model-free reinforcement learning methods are based on random sampling, which severely increases the training consumption and undermines the exploration efficiency. In this letter, we propose an efficient residual policy learning method for high-speed autonomous racing named ResRace, which leverages only the real-time raw observation of LiDAR and IMU for low-latency obstacle avoiding and navigation. We first design a controller based on the modified artificial potential field (MAPF) to generate a policy for navigation. Besides, we utilize the deep reinforcement learning (DRL) algorithm to generate a residual policy as a supplement to obtain the optimal policy. Concurrently, the MAPF policy effectively guides the exploration and increases the update efficiency. This complementary property contributes to the fast convergence and few required resources of our method. We also provide extensive experiments to illustrate our method outperforms the leading algorithms and reaches the comparable level of professional human players on the five F1Tenth tracks.
引用
收藏
页码:11625 / 11632
页数:8
相关论文
共 50 条
[1]   An Online Training Method for Augmenting MPC with Deep Reinforcement Learning [J].
Bellegarda, Guillaume ;
Byl, Katie .
2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, :5453-5459
[2]  
Berducci L., 2020, RACECAR GYM
[3]  
Betz J., 2022, ARXIV220207008
[4]  
Bojarski M, 2016, Arxiv, DOI [arXiv:1604.07316, DOI 10.48550/ARXIV.1604.07316]
[5]   Latent Imagination Facilitates Zero-Shot Transfer in Autonomous Racing [J].
Brunnbauer, Axel ;
Berducci, Luigi ;
Brandstatter, Andreas ;
Lechner, Mathias ;
Hasani, Ramin ;
Rus, Daniela ;
Grosu, Radu .
2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA 2022, 2022, :7513-7520
[6]   Event-Based Neuromorphic Vision for Autonomous Driving: A Paradigm Shift for Bio-Inspired Visual Sensing and Perception [J].
Chen, Guang ;
Cao, Hu ;
Conradt, Jorg ;
Tang, Huajin ;
Rohrbein, Florian ;
Knoll, Alois .
IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (04) :34-49
[7]  
Coumans E., 2016, Pybullet, a Python module for physics simulation for games, robotics and machine learning
[8]   Recent Achievements in Model Predictive Control Techniques for Industrial Motor: A Comprehensive State-of-the-Art [J].
Elmorshedy, Mahmoud F. ;
Xu, Wei ;
El-Sousy, Fayez F. M. ;
Islam, Md. Rabiul ;
Ahmed, Abdelsalam A. .
IEEE ACCESS, 2021, 9 :58170-58191
[9]  
Wang YE, 2019, Arxiv, DOI arXiv:1907.10701
[10]   Super-Human Performance in Gran Turismo Sport Using Deep Reinforcement Learning [J].
Fuchs, Florian ;
Song, Yunlong ;
Kaufmann, Elia ;
Scaramuzza, Davide ;
Durr, Peter .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (03) :4257-4264