A deep reinforcement learning based distributed control strategy for connected automated vehicles in mixed traffic platoon

被引:43
作者
Shi, Haotian
Chen, Danjue [1 ,2 ]
Zheng, Nan [3 ]
Wang, Xin [4 ]
Zhou, Yang [5 ]
Ran, Bin [1 ]
机构
[1] Univ Wisconsin, Dept Civil & Environm Engn, 1208 Engn Hall,1415 Engn Dr, Madison, WI 53706 USA
[2] Univ Massachusetts Lowell, Dept Civil & Environm Engn, 1 Univ Ave,Shal Hall 2000, Lowell, MA 01854 USA
[3] Monash Univ, Dept Civil Engn, 23 Coll Walk B60,Clayton Campus, Monash Clayton, Vic 3800, Australia
[4] Dept Ind Syst Engn, 3258 Mech Engn Bldg 1513 Univ Ave, Madison, WI 53706 USA
[5] Texas A&M Univ, Zachry Dept Civil & Environm Engn, 199 Spence St,DLEB 301,3136 TAMU, College Stn, TX 77843 USA
关键词
Mixed traffic environment; Distributed control; Deep reinforcement learning; Traffic oscillation dampening; Connected automated vehicle; MODEL-PREDICTIVE CONTROL; ROLLING HORIZON CONTROL; CAR-FOLLOWING MODEL; STRING STABILITY; SYSTEMS; VALIDATION; FRAMEWORK; WAVELET; IMPACT; CACC;
D O I
10.1016/j.trc.2023.104019
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
This paper proposes an innovative distributed longitudinal control strategy for connected auto-mated vehicles (CAVs) in the mixed traffic environment of CAV and human-driven vehicles (HDVs), incorporating high-dimensional platoon information. For mixed traffic, the traditional CAV control method focuses on microscopic trajectory information, which may not be efficient in handling the HDV stochasticity (e.g., long reaction time; various driving styles) and mixed traffic heterogeneities. Different from traditional methods, our method, for the first time, characterizes consecutive HDVs as a whole (i.e., AHDV) to reduce the HDV stochasticity and utilize its macroscopic features to control the following CAVs. The new control strategy takes advantage of platoon information to anticipate the disturbances and traffic features induced downstream under mixed traffic scenarios and greatly outperforms the traditional methods. In particular, the control algorithm is based on deep reinforcement learning (DRL) to fulfill car-following control efficiency and further address the stochasticity for the aggregated car following behavior by embedding it in the training environment. To better utilize the macroscopic traffic features, a general platoon of mixed traffic is categorized as a CAV-HDVs-CAV pattern and described by corresponding DRL states. The macroscopic traffic flow properties are built upon the Newell car-following model to capture the characteristics of aggregated HDVs' joint behaviors. Simulated experiments are conducted to validate our proposed strategy. The results demonstrate that the proposed control method has outstanding performances in terms of oscillation dampening, eco-driving, and generalization capability.
引用
收藏
页数:27
相关论文
共 72 条
[1]   On the stable equilibrium points of gradient systems [J].
Absil, P-A. ;
Kurdyka, K. .
SYSTEMS & CONTROL LETTERS, 2006, 55 (07) :573-577
[2]  
[Anonymous], 2016, PROC 4 INT C LEARN R
[3]   Mixed Traffic of Connected and Autonomous Vehicles and Human-Driven Vehicles: Traffic Evolution and Control using Spring-Mass-Damper System [J].
Bang, Soohyuk ;
Ahn, Soyoung .
TRANSPORTATION RESEARCH RECORD, 2019, 2673 (07) :504-515
[4]   Ecological control strategy for cooperative autonomous vehicle in mixed traffic considering linear stability [J].
Lu C. ;
Liu C. .
Journal of Intelligent and Connected Vehicles, 2021, 4 (03) :115-124
[5]   Traffic dynamics under speed disturbance in mixed traffic with automated and non-automated vehicles [J].
Chen, Danjue ;
Srivastava, Anupam ;
Ahn, Soyoung ;
Li, Tienan .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2020, 113 (113) :293-313
[6]   A behavioral car-following model that captures traffic oscillations [J].
Chen, Danjue ;
Laval, Jorge ;
Zheng, Zuduo ;
Ahn, Soyoung .
TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2012, 46 (06) :744-761
[7]   Information Dissemination Delay in Vehicle-to-Vehicle Communication Networks in a Traffic Stream [J].
Du, Lili ;
Dao, Hoang .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2015, 16 (01) :66-80
[8]  
Du RJ, 2020, Arxiv, DOI arXiv:2010.05439
[9]   Hierarchical reinforcement learning for self-driving decision-making without reliance on labelled driving data [J].
Duan, Jingliang ;
Eben Li, Shengbo ;
Guan, Yang ;
Sun, Qi ;
Cheng, Bo .
IET INTELLIGENT TRANSPORT SYSTEMS, 2020, 14 (05) :297-305
[10]   Passing Rates to Measure Relaxation and Impact of Lane-Changing in Congestion [J].
Duret, Aurelien ;
Ahn, Soyoung ;
Buisson, Christine .
COMPUTER-AIDED CIVIL AND INFRASTRUCTURE ENGINEERING, 2011, 26 (04) :285-297