Self-adaptive Torque Vectoring Controller Using Reinforcement Learning

被引:3
作者
Taherian, Shayan [1 ]
Kuutti, Sampo [1 ]
Visca, Marco [1 ]
Fallah, Saber [1 ]
机构
[1] Univ Surrey, CAV Lab, Dept Mech Engn, Guildford, Surrey, England
来源
2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC) | 2021年
关键词
NEURAL-NETWORKS;
D O I
10.1109/ITSC48978.2021.9564494
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continuous direct yaw moment control systems such as torque-vectoring controller are an essential part for vehicle stabilization. This controller has been extensively researched with the central objective of maintaining the vehicle stability by providing consistent stable cornering response. The ability of careful tuning of the parameters in a torque-vectoring controller can significantly enhance vehicle's performance and stability. However, without any re-tuning of the parameters, especially in extreme driving conditions e.g. low friction surface or high velocity, the vehicle fails to maintain the stability. In this paper, the utility of Reinforcement Learning (RL) based on Deep Deterministic Policy Gradient (DDPG) as a parameter tuning algorithm for torque-vectoring controller is presented. It is shown that, torque-vectoring controller with parameter tuning via reinforcement learning performs well on a range of different driving environment e.g., wide range of friction conditions and different velocities, which highlight the advantages of reinforcement learning as an adaptive algorithm for parameter tuning. Moreover, the robustness of DDPG algorithm are validated under scenarios which are beyond the training environment of the reinforcement learning algorithm. The simulation has been carried out using a four wheels vehicle model with nonlinear tire characteristics. We compare our DDPG based parameter tuning against a genetic algorithm and a conventional trial-and-error tunning of the torque vectoring controller, and the results demonstrated that the reinforcement learning based parameter tuning significantly improves the stability of the vehicle.
引用
收藏
页码:172 / 179
页数:8
相关论文
共 24 条
[1]  
[Anonymous], 2012, ISRN CHEM ENG, DOI DOI 10.1155/2012/329389
[2]   PID controller optimized by genetic algorithm for direct-drive servo system [J].
Cao, Fulu .
NEURAL COMPUTING & APPLICATIONS, 2020, 32 (01) :23-30
[3]   Adaptive Fuzzy Sliding Mode Control for Nonlinear Uncertain SISO System Optimized by Differential Evolution Algorithm [J].
Cao Van Kien ;
Nguyen Ngoc Son ;
Ho Pham Huy Anh .
INTERNATIONAL JOURNAL OF FUZZY SYSTEMS, 2019, 21 (03) :755-768
[4]  
Chen JH, 2004, J PROCESS CONTR, V14, P211, DOI 10.1016/S0959-1524(03)000039-8
[5]  
Falcone P, 2007, IEEE DECIS CONTR P, P800
[6]   Vehicle Optimal Torque Vectoring Using State-Derivative Feedback and Linear Matrix Inequality [J].
Fallah, Saber ;
Khajepour, Amir ;
Fidan, Baris ;
Chen, Shih-Ken ;
Litkouhi, Bakhtiar .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2013, 62 (04) :1540-1552
[7]   Adaptive Sliding Mode Control of Dynamic Systems Using Double Loop Recurrent Neural Network Structure [J].
Fei, Juntao ;
Lu, Cheng .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (04) :1275-1286
[8]  
Fu YX, 2016, PROCEEDINGS OF THE 2016 12TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA), P691, DOI 10.1109/WCICA.2016.7578468
[9]  
Hsu Y., 2011, REV SELECTED REV REI, P1
[10]   On actor-critic algorithms [J].
Konda, VR ;
Tsitsiklis, JN .
SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2003, 42 (04) :1143-1166