Double Deep Q-Network with a Dual-Agent for Traffic Signal Control

被引:25
作者
Gu, Jianfeng [1 ]
Fang, Yong [1 ]
Sheng, Zhichao [1 ]
Wen, Peng [2 ]
机构
[1] Shanghai Univ, Shanghai Inst Adv Commun & Data Sci, Key Lab Specialty Fiber Opt & Opt Access Networks, Joint Int Res Lab Specialty Fiber Opt & Adv Commu, Shanghai 200444, Peoples R China
[2] Univ Southern Queensland, Sch Mech & Elect Engn, Toowoomba, Qld 4350, Australia
来源
APPLIED SCIENCES-BASEL | 2020年 / 10卷 / 05期
基金
中国国家自然科学基金;
关键词
adaptive traffic signal control; deep reinforcement learning; Double Deep Q-Network;
D O I
10.3390/app10051622
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Adaptive traffic signal control (ATSC) based on deep reinforcement learning (DRL) has shown promising prospects to reduce traffic congestion. Most existing methods keeping traffic signal phases fixed adopt two agent actions to match a four-phase suffering unstable performance and undesirable operation in a four-phase signalized intersection. In this paper, a Double Deep Q-Network (DDQN) with a dual-agent algorithm is proposed to obtain a stable traffic signal control policy. Specifically, two agents are denoted by two different states and shift the control of green lights to make the phase sequence fixed and control process stable. State representations and reward functions are presented by improving the observability and reducing the leaning difficulty of two agents. To enhance the feasibility and reliability of two agents in the traffic control of the four-phase signalized intersection, a network structure incorporating DDQN is proposed to map states to rewards. Experiments under Simulation of Urban Mobility (SUMO) are carried out, and results show that the proposed traffic signal control algorithm is effective in improving traffic capacity.
引用
收藏
页数:17
相关论文
共 26 条
  • [11] A tutorial survey on vehicular ad hoc networks
    Hartenstein, Hannes
    Laberteaux, Kenneth P.
    [J]. IEEE COMMUNICATIONS MAGAZINE, 2008, 46 (06) : 164 - 171
  • [12] Hausknecht M., 2015, P INT C ART INT AAAI
  • [13] An integrated simulation-based fuzzy logic model for real-time traffic signal control
    Hawas, Y. E.
    [J]. TRANSPORTMETRICA, 2011, 7 (04): : 247 - 278
  • [14] Kingma Diederik P., 2015, 3 INT C LEARNING REP
  • [15] Krajzewicz D., 2012, INT J ADV SYST MEAS, V5, P3
  • [16] Lee J., 2005, Journal of Intelligent Transportation Systems, V9, P111, DOI [10.1080/15472450500183649, DOI 10.1080/15472450500183649]
  • [17] Traffic signal timing via deep reinforcement learning
    Li L.
    Lv Y.
    Wang F.-Y.
    [J]. Li, Li (li-li@tsinghua.edu.cn), 1600, Institute of Electrical and Electronics Engineers Inc. (03): : 247 - 254
  • [18] A Deep Reinforcement Learning Network for Traffic Light Cycle Control
    Liang, Xiaoyuan
    Du, Xunsheng
    Wang, Guiling
    Han, Zhu
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (02) : 1243 - 1253
  • [19] Distributed Cooperative Reinforcement Learning-Based Traffic Signal Control That Integrates V2X Networks' Dynamic Clustering
    Liu, Weirong
    Qin, Gaorong
    He, Yun
    Jiang, Fei
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (10) : 8667 - 8681
  • [20] Human-level control through deep reinforcement learning
    Mnih, Volodymyr
    Kavukcuoglu, Koray
    Silver, David
    Rusu, Andrei A.
    Veness, Joel
    Bellemare, Marc G.
    Graves, Alex
    Riedmiller, Martin
    Fidjeland, Andreas K.
    Ostrovski, Georg
    Petersen, Stig
    Beattie, Charles
    Sadik, Amir
    Antonoglou, Ioannis
    King, Helen
    Kumaran, Dharshan
    Wierstra, Daan
    Legg, Shane
    Hassabis, Demis
    [J]. NATURE, 2015, 518 (7540) : 529 - 533