Deep reinforcement learning-based multi-objective control of hybrid power system combined with road recognition under time-varying environment

被引:45
作者
Chen, Jiaxin [1 ]
Shu, Hong [1 ]
Tang, Xiaolin [1 ]
Liu, Teng [2 ]
Wang, Weida [3 ]
机构
[1] Chongqing Univ, Coll Mech & Vehicle Engn, Chongqing 400044, Peoples R China
[2] Univ Waterloo, Dept Mech & Mech Engn, Waterloo, ON N2L 3G1, Canada
[3] Beijing Inst Technol, Sch Mech Engn, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Hybrid electric vehicle; Road recognition network; Deep reinforcement learning; Multi-objective control network; Energy management strategy; ENERGY MANAGEMENT; ELECTRIC VEHICLES; MODEL; STRATEGY;
D O I
10.1016/j.energy.2021.122123
中图分类号
O414.1 [热力学];
学科分类号
摘要
Aiming at promoting the intelligent development of control technology for new energy vehicles and showing the outstanding advantages of deep reinforcement learning (DRL), this paper trained a VGG16based road recognition convolutional neural network (CNN) at first. Lots of high-definition images of five typical roads are collected by the racing game Dust Rally 2.0, including dry asphalt, wet asphalt, snow, dry cobblestone, and wet cobblestone. Then, a time-varying driving environment model was established, involving driving images, road slope, longitudinal speed, and the number of passengers. Finally, a stereoscopic control network suitable for nine-dimensional state space and three-dimensional action space was built, and for parallel hybrid electric vehicles (HEVs) with the P3 structure, a deep q-network (DQN)based energy management strategy (EMS) achieving multi-objective control was proposed, including the fine-tuning strategy of the motor speed to maintain the optimal slip rate during braking, the engine power control strategy and the continuously variable transmission (CVT) gear ratio control strategy. Simulation results show under the influence of some factors such as tree shade and image compression, the road recognition network has the highest accuracy for snow roads and wet asphalt roads. Three types of control strategies learned simultaneously by the stereoscopic control network not only maintain the near-optimal slip rate in the braking period but also achieve a fuel consumption of 4788.93 g, while dynamic programming (DP)-based EMS gets a fuel consumption of 4295.61 g. Moreover, even DP-based EMS only contains three states and two actions, the time consumed for DP-based EMS and DQN-based EMS to run the speed cycle of 3602s is about 4911s and 10s, respectively. Therefore, the optimization and real-time performance of DRL-based EMS can be guaranteed. (c) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:15
相关论文
共 43 条
[1]  
Burckhardt M., 1993, Fahrwerktechnik: radschlupf-regelsysteme, V36
[2]   A geometry-driven car-following distance estimation algorithm robust to road slopes [J].
Cao, Zhong ;
Yang, Diange ;
Jiang, Kun ;
Xu, Shaobing ;
Wang, Sijia ;
Zhu, Minghan ;
Xiao, Zhongyang .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2019, 102 :274-288
[3]   Thorough state-of-the-art analysis of electric and hybrid vehicle powertrains: Topologies and integrated energy management strategies [J].
Dai-Duong Tran ;
Vafaeipour, Majid ;
El Baghdadi, Mohamed ;
Barrero, Ricardo ;
Van Mierlo, Joeri ;
Hegazy, Omar .
RENEWABLE & SUSTAINABLE ENERGY REVIEWS, 2020, 119
[4]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[5]   Model predictive control of hybrid electric vehicles for fuel economy, emission reductions, and inter-vehicle safety in car-following scenarios [J].
Hu, Xiaosong ;
Zhang, Xiaoqian ;
Tang, Xiaolin ;
Lin, Xianke .
ENERGY, 2020, 196
[6]  
Jazar R.N., 2008, Vehicle Dynamics. Theory and Application
[7]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[8]   Cloud-based health-conscious energy management of hybrid battery systems in electric vehicles with deep reinforcement learning [J].
Li, Weihan ;
Cui, Han ;
Nemeth, Thomas ;
Jansen, Jonathan ;
Uenluebayir, Cem ;
Wei, Zhongbao ;
Feng, Xuning ;
Han, Xuebing ;
Ouyang, Minggao ;
Dai, Haifeng ;
Wei, Xuezhe ;
Sauer, Dirk Uwe .
APPLIED ENERGY, 2021, 293
[9]   Energy management for a power-split hybrid electric bus via deep reinforcement learning with terrain information [J].
Li, Yuecheng ;
He, Hongwen ;
Khajepour, Amir ;
Wang, Hong ;
Peng, Jiankun .
APPLIED ENERGY, 2019, 255
[10]   Rule-interposing deep reinforcement learning based energy management strategy for power-split hybrid electric vehicle [J].
Lian, Renzong ;
Peng, Jiankun ;
Wu, Yuankai ;
Tan, Huachun ;
Zhang, Hailong .
ENERGY, 2020, 197