Research on ATO Control Method for Urban Rail Based on Deep Reinforcement Learning

被引:9
作者
Chen, Xiaoqiang [1 ,2 ,3 ,4 ]
Guo, Xiao [1 ]
Meng, Jianjun [1 ,2 ,3 ]
Xu, Ruxun [1 ,2 ,3 ,4 ]
Li, Shanshan [1 ]
Li, Decang [1 ,2 ,3 ]
机构
[1] Lanzhou Jiaotong Univ, Mechatron T&R Inst, Lanzhou 730070, Peoples R China
[2] Gansu Logist & Transportat Equipment Informat Tech, Lanzhou 730070, Peoples R China
[3] Gansu Logist & Transportat Equipment Ind Tech Ctr, Lanzhou 730070, Peoples R China
[4] Lanzhou Jiaotong Univ, Sch Mech Engn, Lanzhou 730070, Peoples R China
基金
中国国家自然科学基金;
关键词
Energy consumption; Railway transportation; Reinforcement learning; Heuristic algorithms; Target tracking; Real-time systems; Resource management; Urban areas; Public transportation; Urban rail train; DQN algorithm; multi-objective optimization; automatic driving; TRAIN OPERATION; OPTIMIZATION; NETWORK; SYSTEM; SUBWAY;
D O I
10.1109/ACCESS.2023.3236413
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Aiming at the problems of punctuality, parking accuracy and energy saving of urban rail train operation, an intelligent control method for automatic train operation (ATO) based on deep Q network (DQN) is proposed. The train dynamics model is established under the condition of satisfying the safety principle and various constraints of automatic driving of urban rail train. Considering the transformation rules and sequences of working conditions between train stations, the agent in the DQN algorithm is used as the train controller to adjust the train automatic driving strategy in real time according to the train operating state and operating environment, and optimizes the generation of the train automatic driving curve. Taking the Beijing Yizhuang Subway line as an example, the simulation test results show that the DQN urban rail train control method reduces energy consumption by 12.32% compared with the traditional train PID control method, and improves the running punctuality and parking accuracy; at the same time, the DQN train automatically driving control method can adjust the train running state in real time and dynamically, and has good adaptability and robustness to the change of train running environment parameters.
引用
收藏
页码:5919 / 5928
页数:10
相关论文
共 23 条
[1]   Online Learning Algorithms for Train Automatic Stop Control Using Precise Location Data of Balises [J].
Chen, Dewang ;
Chen, Rong ;
Li, Yidong ;
Tang, Tao .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2013, 14 (03) :1526-1535
[2]  
[董昱 Dong Yu], 2021, [电子测量与仪器学报, Journal of Electronic Measurement and Instrument], V35, P103
[3]  
Hou X.F., 2021, URBAN RAPID RAIL TRA, V34, P12, DOI 10.3969/j.issn.1672-6073.2021.01.003
[4]  
Huang You-neng, 2016, Journal of Traffic and Transportation Engineering, V16, P118
[5]  
[李诚 Li Cheng], 2017, [铁道学报, Journal of the China Railway Society], V39, P53
[6]   PID Control to Maglev Train System [J].
Liu, Hengkun ;
Zhang, Xiao ;
Chang, Wensen .
2009 INTERNATIONAL CONFERENCE ON INDUSTRIAL AND INFORMATION SYSTEMS, PROCEEDINGS, 2009, :341-343
[7]   Intelligent Traction Control Method Based on Model Predictive Fuzzy PID Control and Online Optimization for Permanent Magnetic Maglev Trains [J].
Liu, Yahui ;
Fan, Kuangang ;
Ouyang, Qinghua .
IEEE ACCESS, 2021, 9 :29032-29046
[8]  
Mnih V, 2013, Arxiv, DOI [arXiv:1312.5602, 10.48550/arXiv.1312.5602]
[9]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[10]   A review of methods to measure and calculate train resistances [J].
Rochard, BP ;
Schmid, F .
PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART F-JOURNAL OF RAIL AND RAPID TRANSIT, 2000, 214 (04) :185-199