Deep Reinforcement Learning Based PHEV Energy Management With Co-Recognition for Traffic Condition and Driving Style

被引:31
作者
Cui, Naxin [1 ]
Cui, Wei [1 ]
Shi, Yuemei [1 ]
机构
[1] Shandong Univ, Sch Control Sci & Engn, Jinan 250061, Peoples R China
来源
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES | 2023年 / 8卷 / 04期
基金
中国国家自然科学基金;
关键词
Intelligent vehicles; Driving style recognition; energy management strategy; plug-in hybrid electric vehicle; traffic condition recognition; twin delayed deep deterministic policy gradient; RECENT PROGRESS; STRATEGIES;
D O I
10.1109/TIV.2023.3235110
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As one of the key technologies of plug-in hybrid electric vehicle (PHEV), the energy management strategy (EMS) is interactively influenced by driving style, traffic condition as well as vehicle operation state. In this paper, a novel twin delayed deep deterministic police gradient (TD3) algorithm based EMS integrating co-recognition for driving style and traffic condition is proposed, aiming to improve the generalization ability of EMS in various application scenarios with superior energy saving performance and higher self-learning efficiency. In particular, the TD3 based energy management architecture combining delayed policy update and smooth regularization technologies is studied to achieve simultaneous improvement for PHEV energy efficiency and strategy convergence speed. Furthermore, the traffic conditions are recognized by fuzzy C-means method, while the local minimum problem is effectively avoided by incorporating simulated annealing (SA) and genetic algorithm (GA). Sequentially, the driving styles are decoupled from recognized traffic condition, which are further recognized as three typical styles. The comparison results of the proposed strategy with several representative deep reinforcement learning based EMSs indicate that the TD3 based EMS outperforms DDQN and DDPG based EMSs in terms of convergence speed and energy saving performance. With considering the recognized traffic condition and driving style, the energy efficiency of TD3 based EMS is further improved with an ideal robustness to various driving cycles.
引用
收藏
页码:3026 / 3039
页数:14
相关论文
共 43 条
[1]   Efficacy of Haptic Pedal Feel Compensation on Driving With Regenerative Braking [J].
Caliskan, Umut ;
Patoglu, Volkan .
IEEE TRANSACTIONS ON HAPTICS, 2020, 13 (01) :175-182
[2]   An efficient multi-objective hierarchical energy management strategy for plug-in hybrid electric vehicle in connected scenario [J].
Cui, Wei ;
Cui, Naxin ;
Li, Tao ;
Cui, Zhongrui ;
Du, Yi ;
Zhang, Chenghui .
ENERGY, 2022, 257
[3]  
Dörr D, 2014, 2014 IEEE 17TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), P1021, DOI 10.1109/ITSC.2014.6957822
[4]  
Dong H., 2020, Deep Reinforcement Learning: Fundamentals, Research and Applications
[5]   Engine ON/OFF Control for the Energy Management of a Serial Hybrid Electric Bus via Convex Optimization [J].
Elbert, Philipp ;
Nueesch, Tobias ;
Ritter, Andreas ;
Murgovski, Nikolce ;
Guzzella, Lino .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2014, 63 (08) :3549-3559
[6]  
Fujimoto S, 2018, Arxiv, DOI [arXiv:1802.09477, 10.48550/arXiv.1802.09477]
[7]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[8]   A Computationally Efficient and Hierarchical Control Strategy for Velocity Optimization of On-Road Vehicles [J].
Guo, Lulu ;
Chen, Hong ;
Liu, Qifang ;
Gao, Bingzhao .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2019, 49 (01) :31-41
[9]   Adaptive optimal control based on driving style recognition for plug-in hybrid electric vehicle [J].
Guo, Qiuyi ;
Zhao, Zhiguo ;
Shen, Peihong ;
Zhan, Xiaowen ;
Li, Jingwei .
ENERGY, 2019, 186
[10]   Energy management based on reinforcement learning with double deep Q-learning for a hybrid electric tracked vehicle [J].
Han, Xuefeng ;
He, Hongwen ;
Wu, Jingda ;
Peng, Jiankun ;
Li, Yuecheng .
APPLIED ENERGY, 2019, 254