Deep reinforcement learning-based multi-objective control of hybrid power system combined with road recognition under time-varying environment

被引:41
作者
Chen, Jiaxin [1 ]
Shu, Hong [1 ]
Tang, Xiaolin [1 ]
Liu, Teng [2 ]
Wang, Weida [3 ]
机构
[1] Chongqing Univ, Coll Mech & Vehicle Engn, Chongqing 400044, Peoples R China
[2] Univ Waterloo, Dept Mech & Mech Engn, Waterloo, ON N2L 3G1, Canada
[3] Beijing Inst Technol, Sch Mech Engn, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Hybrid electric vehicle; Road recognition network; Deep reinforcement learning; Multi-objective control network; Energy management strategy; ENERGY MANAGEMENT; ELECTRIC VEHICLES; MODEL; STRATEGY;
D O I
10.1016/j.energy.2021.122123
中图分类号
O414.1 [热力学];
学科分类号
摘要
Aiming at promoting the intelligent development of control technology for new energy vehicles and showing the outstanding advantages of deep reinforcement learning (DRL), this paper trained a VGG16based road recognition convolutional neural network (CNN) at first. Lots of high-definition images of five typical roads are collected by the racing game Dust Rally 2.0, including dry asphalt, wet asphalt, snow, dry cobblestone, and wet cobblestone. Then, a time-varying driving environment model was established, involving driving images, road slope, longitudinal speed, and the number of passengers. Finally, a stereoscopic control network suitable for nine-dimensional state space and three-dimensional action space was built, and for parallel hybrid electric vehicles (HEVs) with the P3 structure, a deep q-network (DQN)based energy management strategy (EMS) achieving multi-objective control was proposed, including the fine-tuning strategy of the motor speed to maintain the optimal slip rate during braking, the engine power control strategy and the continuously variable transmission (CVT) gear ratio control strategy. Simulation results show under the influence of some factors such as tree shade and image compression, the road recognition network has the highest accuracy for snow roads and wet asphalt roads. Three types of control strategies learned simultaneously by the stereoscopic control network not only maintain the near-optimal slip rate in the braking period but also achieve a fuel consumption of 4788.93 g, while dynamic programming (DP)-based EMS gets a fuel consumption of 4295.61 g. Moreover, even DP-based EMS only contains three states and two actions, the time consumed for DP-based EMS and DQN-based EMS to run the speed cycle of 3602s is about 4911s and 10s, respectively. Therefore, the optimization and real-time performance of DRL-based EMS can be guaranteed. (c) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页数:15
相关论文
共 43 条
  • [1] Burckhardt M., 1993, FAHRWERKTECHNIK RADS
  • [2] A geometry-driven car-following distance estimation algorithm robust to road slopes
    Cao, Zhong
    Yang, Diange
    Jiang, Kun
    Xu, Shaobing
    Wang, Sijia
    Zhu, Minghan
    Xiao, Zhongyang
    [J]. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2019, 102 : 274 - 288
  • [3] Thorough state-of-the-art analysis of electric and hybrid vehicle powertrains: Topologies and integrated energy management strategies
    Dai-Duong Tran
    Vafaeipour, Majid
    El Baghdadi, Mohamed
    Barrero, Ricardo
    Van Mierlo, Joeri
    Hegazy, Omar
    [J]. RENEWABLE & SUSTAINABLE ENERGY REVIEWS, 2020, 119
  • [4] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [5] Model predictive control of hybrid electric vehicles for fuel economy, emission reductions, and inter-vehicle safety in car-following scenarios
    Hu, Xiaosong
    Zhang, Xiaoqian
    Tang, Xiaolin
    Lin, Xianke
    [J]. ENERGY, 2020, 196
  • [6] Jazar R.N., 2008, VEHICLE DYNAMICS THE
  • [7] ImageNet Classification with Deep Convolutional Neural Networks
    Krizhevsky, Alex
    Sutskever, Ilya
    Hinton, Geoffrey E.
    [J]. COMMUNICATIONS OF THE ACM, 2017, 60 (06) : 84 - 90
  • [8] Cloud-based health-conscious energy management of hybrid battery systems in electric vehicles with deep reinforcement learning
    Li, Weihan
    Cui, Han
    Nemeth, Thomas
    Jansen, Jonathan
    Uenluebayir, Cem
    Wei, Zhongbao
    Feng, Xuning
    Han, Xuebing
    Ouyang, Minggao
    Dai, Haifeng
    Wei, Xuezhe
    Sauer, Dirk Uwe
    [J]. APPLIED ENERGY, 2021, 293
  • [9] Energy management for a power-split hybrid electric bus via deep reinforcement learning with terrain information
    Li, Yuecheng
    He, Hongwen
    Khajepour, Amir
    Wang, Hong
    Peng, Jiankun
    [J]. APPLIED ENERGY, 2019, 255
  • [10] Rule-interposing deep reinforcement learning based energy management strategy for power-split hybrid electric vehicle
    Lian, Renzong
    Peng, Jiankun
    Wu, Yuankai
    Tan, Huachun
    Zhang, Hailong
    [J]. ENERGY, 2020, 197