IG-RL: Inductive Graph Reinforcement Learning for Massive-Scale Traffic Signal Control

被引:63
作者
Devailly, Francois-Xavier [1 ]
Larocque, Denis [1 ]
Charlin, Laurent [1 ]
机构
[1] HEC Montreal, Dept Decis Sci, Montreal, PQ H3T 2A7, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Deep reinforcement learning; transfer learning; adaptive traffic signal control; graph neural networks; zero-shot transfer; independent Q-learning; NETWORK;
D O I
10.1109/TITS.2021.3070835
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Scaling adaptive traffic signal control involves dealing with combinatorial state and action spaces. Multi-agent reinforcement learning attempts to address this challenge by distributing control to specialized agents. However, specialization hinders generalization and transferability, and the computational graphs underlying neural-network architectures-dominating in the multi-agent setting-do not offer the flexibility to handle an arbitrary number of entities which changes both between road networks, and over time as vehicles traverse the network. We introduce Inductive Graph Reinforcement Learning (IG-RL) based on graph-convolutional networks which adapts to the structure of any road network, to learn detailed representations of traffic signal controllers and their surroundings. Our decentralized approach enables learning of a transferable-adaptive-trafficsignal-control policy. After being trained on an arbitrary set of road networks, our model can generalize to new road networks and traffic distributions, with no additional training and a constant number of parameters, enabling greater scalability compared to prior methods. Furthermore, our approach can exploit the granularity of available data by capturing the (dynamic) demand at both the lane level and the vehicle level. The proposed method is tested on both road networks and traffic settings never experienced during training. We compare IG-RL to multi-agent reinforcement learning and domain-specific baselines. In both synthetic road networks and in a larger experiment involving the control of the 3,971 traffic signals of Manhattan, we show that different instantiations of IG-RL outperform baselines.
引用
收藏
页码:7496 / 7507
页数:12
相关论文
共 39 条
[1]   Reinforcement learning for True Adaptive traffic signal control [J].
Abdulhai, B ;
Pringle, R ;
Karakoulas, GJ .
JOURNAL OF TRANSPORTATION ENGINEERING, 2003, 129 (03) :278-285
[2]  
Akkaya I., 2019, SOLVING RUBIKS CUBE
[3]   Reinforcement learning-based multi-agent system for network traffic signal control [J].
Arel, I. ;
Liu, C. ;
Urbanik, T. ;
Kohls, A. G. .
IET INTELLIGENT TRANSPORT SYSTEMS, 2010, 4 (02) :128-135
[4]   Adaptive traffic signal control with actor-critic methods in a real-world traffic network with different traffic disruption events [J].
Aslani, Mohammad ;
Mesgari, Mohammad Saadi ;
Wiering, Marco .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2017, 85 :732-752
[5]  
Bakker B, 2010, STUD COMPUT INTELL, V281, P475
[6]   Real-World Carbon Dioxide Impacts of Traffic Congestion [J].
Barth, Matthew ;
Boriboonsomsin, Kanok .
TRANSPORTATION RESEARCH RECORD, 2008, (2058) :163-171
[7]   A MARKOVIAN DECISION PROCESS [J].
BELLMAN, R .
JOURNAL OF MATHEMATICS AND MECHANICS, 1957, 6 (05) :679-684
[9]  
Busoniu L., 2006, 2006 9 INT C CONTROL, P1
[10]  
Chen CC, 2020, AAAI CONF ARTIF INTE, V34, P3414