Distributed Multiagent Deep Reinforcement Learning for Multiline Dynamic Bus Timetable Optimization

被引:17
作者
Yan, Haoyang [1 ,2 ]
Cui, Zhiyong [3 ]
Chen, Xinqiang [4 ]
Ma, Xiaolei [1 ,2 ]
机构
[1] Beihang Univ, Beijing Key Lab Cooperat Vehicle Infrastruct Syst, Sch Transportat Sci & Engn, Beijing 100191, Peoples R China
[2] Beihang Univ, Beijing Adv Innovat Ctr Big Data & Brain Comp, Beijing 100191, Peoples R China
[3] Univ Washington, Dept Civil & Environm Engn, Seattle, WA 98195 USA
[4] Shanghai Maritime Univ, Inst Logist Sci & Engn, Shanghai 201306, Peoples R China
基金
国家重点研发计划;
关键词
Optimization; Costs; Vehicle dynamics; Heuristic algorithms; Games; Informatics; Data models; Bus timetable optimization; deep reinforcement learning (RL); distributed computing; multiagent; public transit; RESOURCE; SYSTEMS; LEVEL;
D O I
10.1109/TII.2022.3158651
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As a primary countermeasure to mitigate traffic congestion and air pollution, promoting public transit has become a global census. Designing a robust and reliable bus timetable is a pivotal step to increase ridership and reduce operating cost for transit authorities. However, most previous studies on bus timetabling rely on historical passenger count and travel time data to generate static schedules, which often yield biased results in these uncertain scenarios, such as demand surge or adverse weather. In addition, acquiring real-time passenger origin/destination from a limited number of running buses is not feasible. This article considers the multiline dynamic bus timetable optimization problem as a Markov decision process model to address the aforementioned issues, and proposes a multiagent deep reinforcement learning framework to ensure effective learning from the imperfect-information game, where the passenger demand and traffic condition are not always known in advance. Moreover, a distributed reinforcement learning algorithm is applied to overcome the limitation of high computational cost and low efficiency. A case study of multiple bus lines in Beijing, China, confirms the effectiveness and efficiency of the proposed model. The results demonstrate that our method outperforms heuristic and state-of-the-art reinforcement learning algorithms by reducing 20.30% of operating and passenger costs compared with actual timetables.
引用
收藏
页码:469 / 479
页数:11
相关论文
共 30 条
[1]  
Ceder A, 2015, PUBLIC TRANSIT PLANNING AND OPERATION: MODELING, PRACTICE AND BEHAVIOR, 2ND EDITION, P1
[2]   OPTIMAL SCHEDULING OF URBAN TRANSIT SYSTEMS USING GENETIC ALGORITHMS [J].
CHAKROBORTY, P ;
DEB, K ;
SUBRAHMANYAM, PS .
JOURNAL OF TRANSPORTATION ENGINEERING-ASCE, 1995, 121 (06) :544-553
[3]   Reinforcement Learning-Based Load Forecasting of Electric Vehicle Charging Station Using Q-Learning Technique [J].
Dabbaghjamanesh, Morteza ;
Moeini, Amirhossein ;
Kavousi-Fard, Abdollah .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (06) :4229-4237
[4]   Optimising Public Bus Transit Networks Using Deep Reinforcement Learning [J].
Darwish, Ahmed ;
Khali, Momen ;
Badawi, Karim .
2020 IEEE 23RD INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2020,
[5]  
Espeholt L, 2018, PR MACH LEARN RES, V80
[6]   Robust timetable optimization for bus lines subject to resource and regulatory constraints [J].
Gkiotsalitis, K. ;
Alesiani, F. .
TRANSPORTATION RESEARCH PART E-LOGISTICS AND TRANSPORTATION REVIEW, 2019, 128 :30-51
[7]   Use of automatic vehicle location and passenger count data to evaluate bus operations - Experience of the Chicago transit authority, Illinois [J].
Hammerle, M ;
Haynes, M ;
McNeil, S .
TRANSIT: BUS, RURAL PUBLIC AND INTERCITY, AND PARATRANSIT, 2005, (1903) :27-34
[8]   Investigating the irregularity of bus routes: highlighting how underlying assumptions of bus models impact the regularity results [J].
Hans, Etienne ;
Chiabaut, Nicolas ;
Leclercq, Ludovic .
JOURNAL OF ADVANCED TRANSPORTATION, 2015, 49 (03) :358-370
[9]   An analytic stochastic model for the transit vehicle holding problem [J].
Hickman, MD .
TRANSPORTATION SCIENCE, 2001, 35 (03) :215-237
[10]  
Horgan D., 2018, PROC INT C LEARN REP