Tuning path tracking controllers for autonomous cars using reinforcement learning

被引:0
作者
Carrasco A.V. [1 ]
Sequeira J.S. [1 ]
机构
[1] Lisbon University, Instituto Superior Técnico, Lisbon
关键词
Autonomous cars; Autonomous driving systems; Dependability; Non-smooth systems; Path tracking; Q-learning; Reinforcement learning;
D O I
10.7717/PEERJ-CS.1550
中图分类号
学科分类号
摘要
This article proposes an adaptable path tracking control system, based on reinforcement learning (RL), for autonomous cars. A four-parameter controller shapes the behaviour of the vehicle to navigate lane changes and roundabouts. The tuning of the tracker uses an 'educated' Q-Learning algorithm to minimize the lateral and steering trajectory errors, this being a key contribution of this article. The CARLA (CAR Learning to Act) simulator was used both for training and testing. The results show the vehicle is able to adapt its behaviour to the different types of reference trajectories, navigating safely with low tracking errors. The use of a robot operating system (ROS) bridge between CARLA and the tracker (i) results in a realistic system, and (ii) simplifies the replacement of CARLA by a real vehicle, as in a hardware-in-the-loop system. Another contribution of this article is the framework for the dependability of the overall architecture based on stability results of non-smooth systems, presented at the end of this article. © Copyright 2023 Vilaçca Carrasco and Silva Sequeira
引用
收藏
相关论文
共 46 条
[1]  
Ahmed S, Petrov M., Trajectory control of mobile robots using type-2 fuzzy-neural PID controller, IFAC Papers Online, 48, 24, pp. 138-143, (2015)
[2]  
Algarni A, Thayananthan V., Autonomous vehicles: the cybersecurity vulnerabilities and countermeasures for big data communication, Symmetry, 14, 12, (2022)
[3]  
Aubin J, Cellina A., Differential inclusions. Comprehensive studies in mathematics, (1984)
[4]  
Avizienis A, Laprie J, Randell B, Landwehr C., Basic concepts and taxonomy of dependable and secure computing, IEEE Transactions on Dependable and Secure Computing, 1, 1, pp. 11-33, (2004)
[5]  
Bojarski M, Del Testa D, Dworakowski D, Firner B, Flepp B, Goyal P, Jacker L, Monfort M, Muller U, Zhang J, Zhang X, Zhao J, Zieba K., End to end learning for self-driving cars, (2016)
[6]  
Bonsall FF., Lectures on some fixed point theorems of functional analysis, (1962)
[7]  
Brunner M, Rosolia U, Gonzales J, Borrelli F., Repetitive learning model predictive control: an autonomous racing example, 2017 IEEE 56th annual conference on decision and control, pp. 2545-2550, (2017)
[8]  
Chen I, Chan C., Deep reinforcement learning based path tracking controller for autonomous vehicle, Proceedings of the Institution of Mechanical Engineers, Part D: Journal of Automobile Engineering, 235, 2-3, pp. 541-551, (2021)
[9]  
Chen L, Chen Y, Yao X, Shan Y, Chen L., An adaptive path tracking controller based on reinforcement learning with urban driving application, Proceedings. 2019 IEEE intelligent vehicles symposium (IV), pp. 2411-2416, (2019)
[10]  
Deichmann J, Ebel E, Heineke K, Heuss R, Kellner M, Steiner F., Autonomous driving's future: convenient and connected, (2023)