Smart Edge-Enabled Traffic Light Control: Improving Reward-Communication Trade-offs with Federated Reinforcement Learning

被引:7
作者
Hudson, Nathaniel [1 ]
Oza, Pratham [2 ]
Khamfroush, Hana [1 ]
Chantem, Thidapat [2 ]
机构
[1] Univ Kentucky, Dept Comp Sci, Lexington, KY 40506 USA
[2] Virginia Tech, Dept Elect & Comp Engn, Blacksburg, VA USA
来源
2022 IEEE INTERNATIONAL CONFERENCE ON SMART COMPUTING (SMARTCOMP 2022) | 2022年
关键词
Smart Traffic; Traffic Light Control; Reinforcement Learning; Edge Computing; Federated Learning; SIGNAL CONTROL;
D O I
10.1109/SMARTCOMP55677.2022.00021
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traffic congestion is a costly phenomenon of everyday life. Reinforcement Learning (RL) is a promising solution due to its applicability to solving complex decision-making problems in highly dynamic environments. To train smart traffic lights using RL, large amounts of data is required. Recent RL-based approaches consider training to occur on some nearby server or a remote cloud server. However, this requires that traffic lights all communicate their raw data to some central location. For large road systems, communication cost can be impractical, particularly if traffic lights collect heavy data (e.g., video, LIDAR). As such, this work pushes training to the traffic lights directly to reduce communication cost. However, completely independent learning can reduce the performance of trained models. As such, this work considers the recent advent of Federated Reinforcement Learning (FedRL) for edge-enabled traffic lights so they can learn from each other's experience by periodically aggregating locally-learned policy network parameters rather than share raw data, hence keeping communication costs low. To do this, we propose the SEAL framework which uses an intersection-agnostic representation to support FedRL across traffic lights controlling heterogeneous intersection types. We then evaluate our FedRL approach against Centralized and Decentralized RL strategies. We compare the reward-communication trade-offs of these strategies. Our results show that FedRL is able to reduce the communication costs associated with Centralized training by 36.24%; while only seeing a 2.11% decrease in average reward (i.e., decreased traffic congestion).
引用
收藏
页码:40 / 47
页数:8
相关论文
共 30 条
[1]   Reinforcement learning for True Adaptive traffic signal control [J].
Abdulhai, B ;
Pringle, R ;
Karakoulas, GJ .
JOURNAL OF TRANSPORTATION ENGINEERING, 2003, 129 (03) :278-285
[2]  
[Anonymous], 2000, P INT C MACH LEARN
[3]  
Anwer M. S., J EMERGING TRENDS CO
[4]   Adaptive traffic signal control with actor-critic methods in a real-world traffic network with different traffic disruption events [J].
Aslani, Mohammad ;
Mesgari, Mohammad Saadi ;
Wiering, Marco .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2017, 85 :732-752
[5]   Urban traffic signal control using reinforcement learning agents [J].
Balaji, P. G. ;
German, X. ;
Srinivasan, D. .
IET INTELLIGENT TRANSPORT SYSTEMS, 2010, 4 (03) :177-188
[6]   A MARKOVIAN DECISION PROCESS [J].
BELLMAN, R .
JOURNAL OF MATHEMATICS AND MECHANICS, 1957, 6 (05) :679-684
[7]   Vehicular communication: a survey [J].
Bhoi, Sourav Kumar ;
Khilar, Pabitra Mohan .
IET NETWORKS, 2014, 3 (03) :204-217
[8]   Traffic signal timing optimisation based on genetic algorithm approach, including drivers' routing [J].
Ceylan, H ;
Bell, MGH .
TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2004, 38 (04) :329-342
[9]  
Chen Yang., 2021, 2021 IEEE 18th Annual Consumer Communications Networking Conference (CCNC), P1
[10]   Multi-Agent Deep Reinforcement Learning for Large-Scale Traffic Signal Control [J].
Chu, Tianshu ;
Wang, Jie ;
Codeca, Lara ;
Li, Zhaojian .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2020, 21 (03) :1086-1095