Learning a robust multiagent driving policy for traffic congestion reduction

被引:3
作者
Zhang, Yulin [1 ,2 ]
Macke, William [2 ]
Cui, Jiaxun [2 ]
Hornstein, Sharon [3 ]
Urieli, Daniel [3 ]
Stone, Peter [2 ,4 ]
机构
[1] Amazon Robot, 300 Riverpk Dr, North Reading, MA 01864 USA
[2] Univ Texas Austin, Dept Comp Sci, 2317 Speedway, Austin, TX 78712 USA
[3] Gen Motors Israel R&D Labs, Herzliyya, Israel
[4] Sony AI, New York, NY USA
基金
美国国家科学基金会;
关键词
Autonomous vehicles; Deep reinforcement learning; Traffic optimization; Multiagent systems; Multiagent reinforcement learning; Flow; MODEL;
D O I
10.1007/s00521-023-09183-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In most modern cities, traffic congestion is one of the most salient societal challenges. Past research has shown that inserting a limited number of autonomous vehicles (AVs) within the traffic flow, with driving policies learned specifically for the purpose of reducing congestion, can significantly improve traffic conditions. However, to date, these AV policies have generally been evaluated under the same limited conditions under which they were trained. On the other hand, to be considered for practical deployment, they must be robust to a wide variety of traffic conditions. This article establishes for the first time that a multiagent driving policy can be trained in such a way that it generalizes to different traffic flows, AV penetration, and road geometries, including on multilane roads. Inspired by our successful results in a high-fidelity microsimulation, this article further contributes a novel extension of the well-known cell transmission model (CTM) that, unlike the past CTMs, is suitable for modeling congestion in traffic networks, and is thus suitable for studying congestion reduction policies such as those considered in this article.
引用
收藏
页数:14
相关论文
共 27 条