A knowledge-guided reinforcement learning method for lateral path tracking

被引:0
作者
Hu, Bo [1 ]
Zhang, Sunan [2 ]
Feng, Yuxiang [3 ]
Li, Bingbing [2 ]
Sun, Hao [4 ]
Chen, Mingyang [4 ]
Zhuang, Weichao [2 ]
Zhang, Yi [1 ]
机构
[1] Chongqing Univ Technol, Key Lab Adv Mfg Technol Automobile Parts, Minist Educ, Chongqing 400054, Peoples R China
[2] Southeast Univ, Sch Mech Engn, Nanjing 211189, Peoples R China
[3] Imperial Coll London, Dept Civil & Environm Engn, London SW7 2AZ, England
[4] UCL, Dept Elect & Elect Engn, London WC1E 6BT, England
关键词
Lateral control; Knowledge-guided; Reinforcement learning; Online fine-tuning; CONTROLLER; VEHICLES; DESIGN;
D O I
10.1016/j.engappai.2024.109588
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Lateral Control algorithms in autonomous vehicles often necessitates an online fine-tuning procedure in the real world. While reinforcement learning (RL) enables vehicles to learn and improve the lateral control performance through repeated trial and error interactions with a dynamic environment, applying RL directly to safety-critical applications in real physical world is challenging because ensuring safety during the learning process remains difficult. To enable safe learning, a promising direction is to make use of previously gathered offline data, which is frequently accessible in engineering applications. In this context, this paper presents a set of knowledge-guided RL algorithms that can not only fully leverage the prior collected offline data without the need of a physics-based simulator, but also allow further online policy improvement in a smooth, safe and efficient manner. To evaluate the effectiveness of the proposed algorithms on a real controller, a hardware-in-the-loop and a miniature vehicle platform are built. Compared with the vanilla RL, behavior cloning and the existing controller, the proposed algorithms realize a closed-loop solution for lateral control problems from offline training to online fine-tuning, making it attractive for future similar RL-based controller to build upon.
引用
收藏
页数:13
相关论文
共 34 条
[1]   Autonomous vehicle control using a kinematic Lyapunov-based technique with LQR-LMI tuning [J].
Alcala, Eugenio ;
Puig, Vicenc ;
Quevedo, Joseba ;
Escobet, Teresa ;
Comasolivas, Ramon .
CONTROL ENGINEERING PRACTICE, 2018, 73 :1-12
[2]   Three Decades of Driver Assistance Systems Review and Future Perspectives [J].
Bengler, Klaus ;
Dietmayer, Klaus ;
Faerber, Berthold ;
Maurer, Markus ;
Stiller, Christoph ;
Winner, Hermann .
IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE, 2014, 6 (04) :6-22
[3]   Personalized Driver/Vehicle Lane Change Models for ADAS [J].
Butakov, Vadim A. ;
Ioannou, Petros .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2015, 64 (10) :4422-4431
[4]   High-Speed Autonomous Drifting With Deep Reinforcement Learning [J].
Cai, Peide ;
Mei, Xiaodong ;
Tai, Lei ;
Sun, Yuxiang ;
Liu, Ming .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (02) :1247-1254
[5]   Model-Predictive-Control-Based Path Tracking Controller of Autonomous Vehicle Considering Parametric Uncertainties and Velocity-Varying [J].
Cheng, Shuo ;
Li, Liang ;
Chen, Xiang ;
Wu, Jian ;
Wang, Hong-da .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2021, 68 (09) :8698-8707
[6]   Receding horizon lateral vehicle control for pure pursuit path tracking [J].
Elbanhawi, M. ;
Simic, M. ;
Jazar, R. .
JOURNAL OF VIBRATION AND CONTROL, 2018, 24 (03) :619-642
[7]  
Fujimoto S, 2021, ADV NEUR IN, V34
[8]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[9]   Robust Lateral Trajectory Following Control of Unmanned Vehicle Based on Model Predictive Control [J].
Gao, Hongbo ;
Kan, Zhen ;
Li, Keqiang .
IEEE-ASME TRANSACTIONS ON MECHATRONICS, 2022, 27 (03) :1278-1287
[10]  
Hu B., 2023, IEEE T TRANSP ELECT