An Efficient Message Dissemination Scheme for Cooperative Drivings via Cooperative Hierarchical Attention Reinforcement Learning

被引:4
作者
Liu, Bingyi [1 ,2 ,3 ]
Han, Weizhen [1 ]
Wang, Enshu [4 ]
Xiong, Shengwu [1 ]
Qiao, Chunming [5 ]
Wang, Jianping [6 ]
机构
[1] Wuhan Univ Technol, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Wuhan Univ Technol, Sanya Sci & Educ Innovat Pk, Sanya 572025, Peoples R China
[3] Wuhan Univ Technol, Chongqing Res Inst, Chongqing 401120, Peoples R China
[4] Soochow Univ, Dept Future Sci & Engn, Suzhou 214998, Peoples R China
[5] SUNY Buffalo, Dept Comp Sci & Engn, Buffalo, NY 14260 USA
[6] City Univ Hong Kong, Dept Comp Sci, Kowloon, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Decision making; Vehicle dynamics; Games; Electronic mail; Collaboration; Time division multiple access; Cooperative driving; multi-agent reinforcement learning; hierarchical reinforcement learning; graph attention network; CONGESTION CONTROL;
D O I
10.1109/TMC.2023.3312220
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A group of connected and autonomous vehicles with common interests can drive in a cooperative manner, namely cooperative driving. In such a networked control system, an efficient message dissemination scheme is critical for cooperative drivings to periodically broadcast their kinetic status, i.e., beacon. However, most existing researches are designed for a simple or specific scenario, e.g., ignoring the impacts of the complex communication environment and emerging hybrid traffic scenarios. Worse still, the inevitable message transmission interference and the limited interaction among vehicles in harsh communication environments seriously hinder cooperation among cooperative drivings and deteriorate the beaconing performance. In this paper, we formulate the decision-making process of cooperative drivings as a Markov game. Furthermore, we propose a cooperative hierarchical attention reinforcement learning (CHA) framework to solve this Markov game. Specifically, the hierarchical structure of CHA leads cooperative drivings to be foresighted. Besides, we integrate each hierarchical level of CHA separately with graph attention networks to incorporate agents' mutual influences in the decision-making process. Moreover, each hierarchical level learns a cooperative reward function to motivate each agent to cooperate with others under harsh communication conditions. Finally, we set up a simulator and conduct extensive experiments to validate the effectiveness of CHA.
引用
收藏
页码:5527 / 5542
页数:16
相关论文
共 46 条
  • [1] Ahizoune A, 2010, C LOCAL COMPUT NETW, P48, DOI 10.1109/LCN.2010.5735764
  • [2] [Anonymous], 2015, Reinforcement Learning: An Introduction
  • [3] Arjona-Medina JA, 2019, ADV NEUR IN, V32
  • [4] Aslam A, 2017, 17TH IEEE INTERNATIONAL CONFERENCE ON SMART TECHNOLOGIES - IEEE EUROCON 2017 CONFERENCE PROCEEDINGS, P896, DOI 10.1109/EUROCON.2017.8011240
  • [5] Bacon PL, 2017, AAAI CONF ARTIF INTE, P1726
  • [6] Bohm A., 2016, P 12 INT C WIR MOB C, P1
  • [7] Chen Ziyi, 2022, P MACHINE LEARNING R
  • [8] Implementing Distributed TDMA Using Relative Distance in Vehicular Networks
    Deng, Yafeng
    Kim, Dohyung
    Li, Zhetao
    Choi, Young-June
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (07) : 7295 - 7305
  • [9] Du YL, 2019, ADV NEUR IN, V32
  • [10] Efficient Congestion Control With Information Loss Minimization for Vehicular Ad Hoc Networks
    Fang, Yuanming
    Liu, Xuxun
    Li, Zhetao
    Cui, Lin
    Wei, Kaimin
    Deng, Qingyong
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (03) : 3879 - 3888