Powertrain Control for Hybrid-Electric Vehicles Using Supervised Machine Learning

被引:16
作者
Harold, Craig K. D. [1 ]
Prakash, Suraj [1 ]
Hofman, Theo [1 ]
机构
[1] Eindhoven Univ Technol, Mech Engn Dept, Control Syst Technol Grp, NL-5612 AZ Eindhoven, Netherlands
来源
VEHICLES | 2020年 / 2卷 / 02期
关键词
machine learning; powertrain control; automatic re-training; hybrid electric vehicles; dynamic programming; transmission; energy management; ENERGY MANAGEMENT; CONTROL STRATEGY; ECMS; SYSTEM;
D O I
10.3390/vehicles2020015
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
This paper presents a novel framework to enable automatic re-training of the supervisory powertrain control strategy for hybrid electric vehicles using supervised machine learning. The aim of re-training is to customize the control strategy to a user-specific driving behavior without human intervention. The framework is designed to update the control strategy at the end of a driving task. A combination of dynamic programming and supervised machine learning is used to train the control strategy. The trained control strategy denoted as SML is compared to an online-implementable strategy based on the combination of the optimal operation line and Pontryagin's minimum principle denoted as OOL-PMP, on the basis of fuel consumption. SML consistently performed better than OOL-PMP, evaluated over five standard drive cycles. The EUDC performance was almost identical while on FTP75 the OOL-PMP consumed 14.7% more fuel than SML. Moreover, the deviation from the global benchmark obtained from dynamic programming was between 1.8% and 5.4% for SML and between 5.8% and 16.8% for OOL-PMP. Furthermore, a test-case was conducted to emulate a real-world driving scenario wherein a trained controller is exposed to a new drive cycle. It is found that the performance on the new drive cycle deviates significantly from the optimal policy; however, this performance gap is bridged with a single re-training episode for the respective test-case.
引用
收藏
页码:267 / 286
页数:20
相关论文
共 43 条
[1]  
[Anonymous], 2016, Asynchronous methods for deep reinforcement learning
[2]   DYNAMIC PROGRAMMING [J].
BELLMAN, R .
SCIENCE, 1966, 153 (3731) :34-&
[3]  
Bennett G., 1992, AIR POLLUTION CONTRO
[4]  
Bertsekas D. P., 1989, Parallel and Distributed Computation: Numer[1]ical Methods, V23
[5]   An assessment of regulated emissions and CO2 emissions from a European light-duty CNG-fueled vehicle in the context of Euro 6 emissions regulations [J].
Bielaczyc, Piotr ;
Woodburn, Joseph ;
Szczotka, Andrzej .
APPLIED ENERGY, 2014, 117 :134-141
[6]   CVT ratio control strategy optimization [J].
Bonsen, B ;
Steinbuch, M ;
Veenhuizen, PA .
2005 IEEE Vehicle Power and Propulsion Conference (VPPC), 2005, :227-231
[7]  
DeJager B, 2013, ADV IND CONTROL, P1, DOI 10.1007/978-1-4471-5076-3
[8]   Independent driving pattern factors and their influence on fuel-use and exhaust emission factors [J].
Ericsson, E .
TRANSPORTATION RESEARCH PART D-TRANSPORT AND ENVIRONMENT, 2001, 6 (05) :325-345
[9]   Understanding charging behaviour of electric vehicle users [J].
Franke, Thomas ;
Krems, Josef F. .
TRANSPORTATION RESEARCH PART F-TRAFFIC PSYCHOLOGY AND BEHAVIOUR, 2013, 21 :75-89
[10]  
Geulen S, 2015, IEEE INTL CONF CONTR, P1722, DOI 10.1109/CCA.2015.7320858