Cooperative traffic signal control for a partially observed vehicular network using multi-agent reinforcement learning

被引:0
作者
Wang, Chong [1 ,2 ,3 ]
Li, Yueqi [1 ,2 ]
Chen, Jiale [1 ,2 ]
Zhang, Jian [4 ]
Xue, Yu [5 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Jiangsu, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Sch Cyber Sci & Engn, Nanjing 210044, Jiangsu, Peoples R China
[3] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, Minist Educ, Changchun 130012, Jilin, Peoples R China
[4] Southeast Univ, Sch Transportat, Nanjing 211189, Jiangsu, Peoples R China
[5] Nanjing Univ Informat Sci & Technol, Sch Software, Nanjing 210044, Jiangsu, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Intelligent transportation system; Smart cities; Vehicular network; Multi-agent reinforcement learning; Traffic signal control; Vehicle-to-infrastructure system; LIGHT CONTROL;
D O I
10.1016/j.engappai.2025.111813
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cooperative traffic signal control aims to alleviate congestion and reduce travel times in urban environments by coordinating traffic lights. However, accurately modeling urban traffic dynamics and managing multiple intersections remain challenging tasks. Although Reinforcement Learning (RL) offers adaptability without requiring predefined traffic models, it faces challenges such as high dimensionality and data dependency, especially in real-world conditions like adverse weather or device malfunctions. In this paper, we propose a novel multi-tiered, multi-agent RL approach, named Multi-Agent Partially Observed Light (MAPOLight), designed to operate in a Vehicle-to-Infrastructure (V2I) environment with a limited number of Connected and Automated Vehicles (CAVs). The upper tier emphasizes agent collaboration by aggregating states and actions into mean values, significantly reducing state dimensionality. The lower tier employs advanced deep RL algorithms for optimization, ensuring both flexibility and scalability. Additionally, we introduce the Observation Correlation Indicator to relate CAV penetration rates (CAV P-Rates) with RL convergence. Simulation results reveal that our method offers stability and superior performance across diverse scenarios compared to traditional approaches. MAPOLight demonstrates robustness with CAV P-Rates above 20% and achieves convergence with a minimum penetration rate of 5%, outperforming existing methods. Moreover, our approach promotes smoother vehicle trajectories and exhibits strong adaptability, enabling rapid congestion relief in respond to unexpected traffic accidents. These results highlight the effectiveness and adaptability of the proposed approach.
引用
收藏
页数:21
相关论文
共 53 条
[1]   On the relationship between congestion and road safety in cities [J].
Albalate, Daniel ;
Fageda, Xavier .
TRANSPORT POLICY, 2021, 105 :145-152
[2]  
[Anonymous], Pearson correlation coefficient
[3]   Traffic signal optimization through discrete and continuous reinforcement learning with robustness analysis in downtown Tehran [J].
Aslani, Mohammad ;
Seipel, Stefan ;
Mesgari, Mohammad Saadi ;
Wiering, Marco .
ADVANCED ENGINEERING INFORMATICS, 2018, 38 :639-655
[4]   A Collaborative Communication-Qmix Approach for Large-scale Networked Traffic Signal Control [J].
Chen, Xiaoyu ;
Xiong, Gang ;
Lv, Yisheng ;
Chen, Yuanyuan ;
Song, Bing ;
Wang, Fei-Yue .
2021 IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE (ITSC), 2021, :3450-3455
[5]  
Chu T., 2020, INT C LEARN REPR
[6]   Applications in Traffic Signal Control: A Distributed Policy Gradient Decomposition Algorithm [J].
Dai, Pengcheng ;
Yu, Wenwu ;
Wang, He ;
Jiang, Jiahui .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (02) :2762-2775
[7]   MARP: A Cooperative Multiagent DRL System for Connected Autonomous Vehicle Platooning [J].
Dai, Shuhong ;
Li, Shike ;
Tang, Haichuan ;
Ning, Xin ;
Fang, Fang ;
Fu, Yunxiao ;
Wang, Qingle ;
Cheng, Long .
IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (20) :32454-32463
[8]  
Dann C, 2022, PR MACH LEARN RES
[9]   IG-RL: Inductive Graph Reinforcement Learning for Massive-Scale Traffic Signal Control [J].
Devailly, Francois-Xavier ;
Larocque, Denis ;
Charlin, Laurent .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (07) :7496-7507
[10]   Decentralized network level adaptive signal control by multi-agent deep reinforcement learning [J].
Gong, Yaobang ;
Abdel-Aty, Mohamed ;
Cai, Qing ;
Rahman, Md Sharikur .
TRANSPORTATION RESEARCH INTERDISCIPLINARY PERSPECTIVES, 2019, 1