Energy Management Strategy for a Hybrid Electric Vehicle Based on Deep Reinforcement Learning

被引:182
作者
Hu, Yue [1 ,2 ,3 ]
Li, Weimin [1 ,3 ,4 ]
Xu, Kun [1 ]
Zahid, Taimoor [1 ,2 ]
Qin, Feiyan [1 ,2 ]
Li, Chenming [5 ]
机构
[1] Chinese Acad Sci, Shenzhen Inst Adv Technol, Shenzhen 518055, Peoples R China
[2] Univ Chinese Acad Sci, Shenzhen Coll Adv Technol, Shenzhen 518055, Peoples R China
[3] Chinese Acad Sci, Jining Inst Adv Technol, Jining 272000, Peoples R China
[4] Chinese Univ Hong Kong, Dept Mech & Automat Engn, Hong Kong 999077, Hong Kong, Peoples R China
[5] Chinese Univ Hong Kong, Dept Elect Engn, Hong Kong 999077, Hong Kong, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2018年 / 8卷 / 02期
基金
中国国家自然科学基金;
关键词
hybrid electric vehicle; energy management strategy; deep reinforcement learning; online learning; OPTIMIZATION;
D O I
10.3390/app8020187
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
An energy management strategy (EMS) is important for hybrid electric vehicles (HEVs) since it plays a decisive role on the performance of the vehicle. However, the variation of future driving conditions deeply influences the effectiveness of the EMS. Most existing EMS methods simply follow predefined rules that are not adaptive to different driving conditions online. Therefore, it is useful that the EMS can learn from the environment or driving cycle. In this paper, a deep reinforcement learning (DRL)-based EMS is designed such that it can learn to select actions directly from the states without any prediction or predefined rules. Furthermore, a DRL-based online learning architecture is presented. It is significant for applying the DRL algorithm in HEV energy management under different driving conditions. Simulation experiments have been conducted using MATLAB and Advanced Vehicle Simulator (ADVISOR) co-simulation. Experimental results validate the effectiveness of the DRL-based EMS compared with the rule-based EMS in terms of fuel economy. The online learning architecture is also proved to be effective. The proposed method ensures the optimality, as well as real-time applicability, in HEVs.
引用
收藏
页数:15
相关论文
共 29 条
[11]   Optimal fuzzy power control and management of fuel cell/battery hybrid vehicles [J].
Li, Chun-Yan ;
Liu, Guo-Ping .
JOURNAL OF POWER SOURCES, 2009, 192 (02) :525-533
[12]   Online learning control for hybrid electric vehicle [J].
Li Weimin ;
Xu Guoqing ;
Xu Yangsheng .
CHINESE JOURNAL OF MECHANICAL ENGINEERING, 2012, 25 (01) :98-106
[13]   Reinforcement Learning-Based Energy Management Strategy for a Hybrid Electric Tracked Vehicle [J].
Liu, Teng ;
Zou, Yuan ;
Liu, Dexing ;
Sun, Fengchun .
ENERGIES, 2015, 8 (07) :7243-7260
[14]   Energy Management in Plug-in Hybrid Electric Vehicles: Recent Progress and a Connected Vehicles Perspective [J].
Martinez, Clara Marina ;
Hu, Xiaosong ;
Cao, Dongpu ;
Velenis, Efstathios ;
Gao, Bo ;
Wellers, Matthias .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (06) :4534-4549
[15]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[16]   Power management optimization of fuel cell/battery hybrid vehicles with experimental validation [J].
Odeim, Farouk ;
Roes, Juergen ;
Wuelbeck, Lars ;
Heinzel, Angelika .
JOURNAL OF POWER SOURCES, 2014, 252 :333-343
[17]   Data-Driven Reinforcement Learning-Based Real-Time Energy Management System for Plug-In Hybrid Electric Vehicles [J].
Qi, Xuewei ;
Wu, Guoyuan ;
Boriboonsomsin, Kanok ;
Barth, Matthew J. ;
Gonder, Jeffrey .
TRANSPORTATION RESEARCH RECORD, 2016, (2572) :1-8
[18]   Stochastic Optimal Control of Parallel Hybrid Electric Vehicles [J].
Qin, Feiyan ;
Xu, Guoqing ;
Hu, Yue ;
Xu, Kun ;
Li, Weimin .
ENERGIES, 2017, 10 (02)
[19]  
Sallab A.E., 2016, End-to-End Deep Reinforcement Learning for Lane Keeping Assist
[20]   Mastering the game of Go with deep neural networks and tree search [J].
Silver, David ;
Huang, Aja ;
Maddison, Chris J. ;
Guez, Arthur ;
Sifre, Laurent ;
van den Driessche, George ;
Schrittwieser, Julian ;
Antonoglou, Ioannis ;
Panneershelvam, Veda ;
Lanctot, Marc ;
Dieleman, Sander ;
Grewe, Dominik ;
Nham, John ;
Kalchbrenner, Nal ;
Sutskever, Ilya ;
Lillicrap, Timothy ;
Leach, Madeleine ;
Kavukcuoglu, Koray ;
Graepel, Thore ;
Hassabis, Demis .
NATURE, 2016, 529 (7587) :484-+