Safe deep reinforcement learning in diesel engine emission control

被引:15
作者
Norouzi, Armin [1 ,2 ]
Shahpouri, Saeid [1 ]
Gordon, David [1 ]
Shahbakhti, Mahdi [1 ]
Koch, Charles Robert [1 ]
机构
[1] Univ Alberta, Dept Mech Engn, Edmonton, AB, Canada
[2] Univ Alberta, Dept Mech Engn, 116 St & 85 Ave, Edmonton, AB T6G 2R3, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Machine learning; deep learning; reinforcement learning; safe learning; iterative learning control; diesel engine; emission control; SLIDING MODE CONTROL; AIR-FUEL RATIO; SPEED CONTROL; STRATEGIES; VEHICLES; SYSTEMS; LOAD; EGR;
D O I
10.1177/09596518231153445
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A deep reinforcement learning application is investigated to control the emissions of a compression ignition diesel engine. The main purpose of this study is to reduce the engine-out nitrogen oxide ( NOx ) emissions and to minimize fuel consumption while tracking a reference engine load. First, a physics-based engine simulation model is developed in GT-Power and calibrated using experimental data. Using this model and a GT-Power/Simulink co-simulation, a deep deterministic policy gradient is developed. To reduce the risk of an unwanted output, a safety filter is added to the deep reinforcement learning. Based on the simulation results, this filter has no effect on the final trained deep reinforcement learning; however, during the training process, it is crucial to enforce constraints on the controller output. The developed safe reinforcement learning is then compared with an iterative learning controller and a deep neural network-based nonlinear model predictive controller. This comparison shows that the safe reinforcement learning is capable of accurately tracking an arbitrary reference input while the iterative learning controller is limited to a repetitive reference. The comparison between the nonlinear model predictive control and reinforcement learning indicates that for this case reinforcement learning is able to learn the optimal control output directly from the experiment without the need for a model. However, to enforce output constraint for safe learning reinforcement learning, a simple model of system is required. In this work, reinforcement learning was able to reduce NO(x )emissions more than the nonlinear model predictive control; however, it suffered from slightly higher error in load tracking and a higher fuel consumption.
引用
收藏
页码:1440 / 1453
页数:14
相关论文
共 66 条
[1]   Modeling, diagnostics, optimization, and control of internal combustion engines via modern machine learning techniques: A review and future directions [J].
Aliramezani, Masoud ;
Koch, Charles Robert ;
Shahbakhti, Mahdi .
PROGRESS IN ENERGY AND COMBUSTION SCIENCE, 2022, 88
[2]  
Alpaydin E., 2020, Introduction to Machine Learning, V4th
[3]   BETTERING OPERATION OF ROBOTS BY LEARNING [J].
ARIMOTO, S ;
KAWAMURA, S ;
MIYAZAKI, F .
JOURNAL OF ROBOTIC SYSTEMS, 1984, 1 (02) :123-140
[4]  
Astrom K. J., 2013, Adaptive Control
[5]  
Basina LNA, 2020, 2020 IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS (CCTA), P94, DOI [10.1109/CCTA41146.2020.9206358, 10.1109/ccta41146.2020.9206358]
[6]   The Arcade Learning Environment: An Evaluation Platform for General Agents [J].
Bellemare, Marc G. ;
Naddaf, Yavar ;
Veness, Joel ;
Bowling, Michael .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2013, 47 :253-279
[7]   A neural network model-based observer for idle speed control of ignition in SI engine [J].
Czarnigowski, Jacek .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2010, 23 (01) :1-7
[8]   An Introduction to Deep Reinforcement Learning [J].
Francois-Lavet, Vincent ;
Henderson, Peter ;
Islam, Riashat ;
Bellemare, Marc G. ;
Pineau, Joelle .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2018, 11 (3-4) :219-354
[9]  
Gandhi D, 2017, IEEE INT C INT ROBOT, P3948, DOI 10.1109/IROS.2017.8206247
[10]   End-to-End Deep Neural Network Based Nonlinear Model Predictive Control: Experimental Implementation on Diesel Engine Emission Control [J].
Gordon, David C. ;
Norouzi, Armin ;
Winkler, Alexander ;
McNally, Jakub ;
Nuss, Eugen ;
Abel, Dirk ;
Shahbakhti, Mahdi ;
Andert, Jakob ;
Koch, Charles R. .
ENERGIES, 2022, 15 (24)