An Adaptive Model-Free Control Method for Metro Train Based on Deep Reinforcement Learning

被引:0
作者
Lai, Wenzhu [1 ,3 ]
Chen, Dewang [2 ,3 ]
Huang, Yunhu [1 ,3 ]
Huang, Benzun [1 ,3 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou, Peoples R China
[2] Fujian Univ Technol, Sch Transportat, Fuzhou, Peoples R China
[3] Fuzhou Univ, Key Lab Intelligent Metro Univ Fujian Prov, Fuzhou 350108, Peoples R China
来源
ADVANCES IN NATURAL COMPUTATION, FUZZY SYSTEMS AND KNOWLEDGE DISCOVERY, ICNC-FSKD 2022 | 2023年 / 153卷
基金
中国国家自然科学基金;
关键词
Intelligent train operation; Model free adaptive control; Deep reinforcement learning; OPTIMIZATION; OPERATION; SYSTEM; SUBWAY;
D O I
10.1007/978-3-031-20738-9_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The current metro train control system has achieved automatic operation, but the degree of intelligence needs to be enhanced. To improve the intelligence of train driving, this paper adopts the proximal policy optimization (PPO) algorithm to study the intelligent train operation (ITO) of metro trains by drawing on the successful application of deep reinforcement learning in games. We propose an adaptive model-free control (MFAC) method for train speed profile tracking, named as intelligent train operation based on PPO (ITOP), and design reinforcement learning policies, actions, and rewards to ensure the accuracy of the train tracking speed profile, passenger comfort, and stopping accuracy. Simulation experiments are conducted using real railroad data from the Yizhuang Line of Beijing Metro (YLBS). The results show that the tracking curve generated by ITOP is highly coincident with the target curve with good parking accuracy and comfort, and responds positively to the changes of the target curve during the operation. This provides a new solution for the intelligent control of trains.
引用
收藏
页码:263 / 273
页数:11
相关论文
共 16 条
[1]   Bio-Inspired Speed Curve Optimization and Sliding Mode Tracking Control for Subway Trains [J].
Cao, Yuan ;
Wang, Zheng-Chao ;
Liu, Feng ;
Li, Peng ;
Xie, Guo .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (07) :6331-6342
[2]  
Florensa C, 2017, PR MACH LEARN RES, V78
[3]  
Fu PC, 2018, CHIN AUTOM CONGR, P2889, DOI 10.1109/CAC.2018.8623438
[4]   Approximation-Based Robust Adaptive Automatic Train Control: An Approach for Actuator Saturation [J].
Gao, Shigen ;
Dong, Hairong ;
Chen, Yao ;
Ning, Bin ;
Chen, Guanrong ;
Yang, Xiaoxia .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2013, 14 (04) :1733-1742
[5]   Energy-Efficient Train Operation in Urban Rail Transit Using Real-Time Traffic Information [J].
Gu, Qing ;
Tang, Tao ;
Cao, Fang ;
Song, Yong-duan .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2014, 15 (03) :1216-1233
[6]   Reinforcement learning: A survey [J].
Kaelbling, LP ;
Littman, ML ;
Moore, AW .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 1996, 4 :237-285
[7]   Optimization of train-speed trajectory and control for mass rapid transit systems [J].
Ke, Bwo-Ren ;
Lin, Chun-Liang ;
Lai, Chi-Wen .
CONTROL ENGINEERING PRACTICE, 2011, 19 (07) :675-687
[8]  
Lillicrap T. P., 2015, arXiv, DOI DOI 10.48550/ARXIV.1509.02971
[9]   A high speed railway control system based on the fuzzy control method [J].
Liu, W. Y. ;
Han, J. G. ;
Lu, X. N. .
EXPERT SYSTEMS WITH APPLICATIONS, 2013, 40 (15) :6115-6124
[10]  
Mnih V, 2013, Arxiv, DOI [arXiv:1312.5602, 10.48550/arXiv.1312.5602, DOI 10.48550/ARXIV.1312.5602]