Deep Reinforcement Learning Based High-level Driving Behavior Decision-making Model in Heterogeneous Traffic

被引:0
作者
Bai, Zhengwei [1 ]
Wei Shangguan [1 ,2 ]
Cai, Baigen [1 ,2 ]
Chai, Linguo [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Sch Elect & Informat Engn, Beijing 100044, Peoples R China
[2] Beijing Jiaotong Univ, State Key Lab Rail Traff Control & Safety, Beijing 100044, Peoples R China
来源
PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC) | 2019年
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; high-level driving behavior; decision making; connected vehicle; heterogeneous traffic;
D O I
10.23919/chicc.2019.8866005
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
High-level driving behavior decision-making is an open-challenging problem for connected vehicle technology, especially in heterogeneous traffic scenarios. In this paper, a deep reinforcement learning based high-level driving behavior decision-making approach is proposed for connected vehicle in heterogeneous traffic situations. The model is composed of three main parts: a data preprocessor that maps hybrid data into a data format called hyper-grid matrix, a two-stream deep neural network that extracts the hidden features, and a deep reinforcement learning network that learns the optimal policy. Moreover, a simulation environment, which includes different heterogeneous traffic scenarios, is built to train and test the proposed method. The results demonstrate that the model has the capability to learn the optimal high-level driving policy such as driving fast through heterogeneous traffic without unnecessary lane changes. Furthermore, two separate models are used to compare with the proposed model, and the performances are analyzed in detail.
引用
收藏
页码:8600 / 8605
页数:6
相关论文
共 16 条
  • [1] Al-Qizwini M, 2017, IEEE INT VEH SYM, P89, DOI 10.1109/IVS.2017.7995703
  • [2] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [3] Bojarski Mariusz, 2016, arXiv
  • [4] Fridman Lex, 2018, ARXIV180102805
  • [5] Hasselt H. V., 2015, Computer Science
  • [6] Kingma DP, 2014, ARXIV
  • [7] Lillicrap TP, 2015, ARXIV150902971
  • [8] Min K, 2018, IEEE INT VEH SYM, P226, DOI 10.1109/IVS.2018.8500645
  • [9] Human-level control through deep reinforcement learning
    Mnih, Volodymyr
    Kavukcuoglu, Koray
    Silver, David
    Rusu, Andrei A.
    Veness, Joel
    Bellemare, Marc G.
    Graves, Alex
    Riedmiller, Martin
    Fidjeland, Andreas K.
    Ostrovski, Georg
    Petersen, Stig
    Beattie, Charles
    Sadik, Amir
    Antonoglou, Ioannis
    King, Helen
    Kumaran, Dharshan
    Wierstra, Daan
    Legg, Shane
    Hassabis, Demis
    [J]. NATURE, 2015, 518 (7540) : 529 - 533
  • [10] Mnih Volodymyr, 2013, CORR