A Deployment-Efficient Energy Management Strategy for Connected Hybrid Electric Vehicle Based on Offline Reinforcement Learning

被引:46
作者
Hu, Bo [1 ,2 ]
Li, Jiaxi [1 ]
机构
[1] Chongqing Univ Technol, Key Lab Adv Mfg Technol Automobile Parts, Minist Educ, Chongqing 400054, Peoples R China
[2] Ningbo Yinzhou DLT Technol Co Ltd, Ningbo 315000, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Energy management; Hybrid electric vehicles; Training; Task analysis; Heuristic algorithms; Batteries; Safety; Connected; deployment-efficient; energy management strategy (EMS); hybrid electric vehicle (HEV); offline reinforcement learning (RL); OPTIMIZATION; HEVS; GAME; GO;
D O I
10.1109/TIE.2021.3116581
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the development of recent artificial intelligence technology, especially after the great success of AlphaGo, there has been a growing interest in applying reinforcement learning (RL) to solve energy management strategy (EMS) problems for hybrid electric vehicles. However, the issues of current RL algorithms including deployment inefficiency, safety constraint, and simulation-to-real gap make it inapplicable to many industrial EMS tasks. With these in mind and considering the fact that there exists many suboptimal EMS controllers which can generate plentiful amounts of interactive data containing informative behaviors, an offline RL training framework that tries to extract policies with the maximum possible utility out of the available offline data is proposed. Furthermore, with connected vehicle technology standard in many new cars, rather than bringing all the data to the storage and analytics, a scheduled training framework is put forward. This cloud-based approach not only alleviates the computational burden of edge devices, but also more importantly provides a deployment-efficient solution to EMS tasks that have to adapt to changes of driving cycle. To evaluate the effectiveness of the proposed algorithm on real controllers, a hardware-in-the-loop (HIL) test is performed and the superiority of the proposed algorithm in contrast to dynamic programming, behavior cloning, rule-based, and vanilla off-policy RL algorithms is given.
引用
收藏
页码:9644 / 9654
页数:11
相关论文
共 33 条
[1]  
Alshiekh M, 2018, AAAI CONF ARTIF INTE, P2669
[2]   Modelling and control of hybrid electric vehicles (A comprehensive review) [J].
Enang, Wisdom ;
Bannister, Chris .
RENEWABLE & SUSTAINABLE ENERGY REVIEWS, 2017, 74 :1210-1239
[3]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[4]  
García J, 2015, J MACH LEARN RES, V16, P1437
[5]   Optimal Energy Management for HEVs in Eco-Driving Applications Using Bi-Level MPC [J].
Guo, Lulu ;
Gao, Bingzhao ;
Gao, Ying ;
Chen, Hong .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2017, 18 (08) :2153-2162
[6]   Shifting Deep Reinforcement Learning Algorithm Toward Training Directly in Transient Real-World Environment: A Case Study in Powertrain Control [J].
Hu, Bo ;
Li, Jiaxi .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (12) :8198-8206
[7]   An Edge Computing Framework for Powertrain Control System Optimization of Intelligent and Connected Vehicles Based on Curiosity-Driven Deep Reinforcement Learning [J].
Hu, Bo ;
Li, Jiaxi .
IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2021, 68 (08) :7652-7661
[8]   Comparative Study of Real-Time HEV Energy Management Strategies [J].
Jiang, Qi ;
Ossart, Florence ;
Marchand, Claude .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (12) :10875-10888
[9]   Optimal Control of Hybrid Electric Vehicles Based on Pontryagin's Minimum Principle [J].
Kim, Namwook ;
Cha, Sukwon ;
Peng, Huei .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2011, 19 (05) :1279-1287
[10]  
Kingma Diederik P, 2013, ICLR 2014