Behavior analysis of emergent rule discovery for cooperative automated driving using deep reinforcement learning

被引:9
作者
Harada, Tomohiro [1 ]
Matsuoka, Johei [2 ]
Hattori, Kiyohiko [2 ]
机构
[1] Tokyo Metropolitan Univ, Fac Syst Design, Tokyo, Japan
[2] Tokyo Univ Technol, Sch Comp Sci, Tokyo, Japan
关键词
Multi-agent learning; Autonomous driving; Reinforcement learning; Reward design; Proximal policy optimization; Vehicle-to-vehicle communication;
D O I
10.1007/s10015-022-00839-7
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
With the improvements in AI technology and sensor performance, research on automated driving has become increasingly popular. However, most studies are based on human driving styles. In this study, we consider an environment in which only autonomous vehicles are present. In such an environment, it is essential to develop an appropriate control method that actively utilizes the characteristics of autonomous vehicles, such as dense information exchange and highly accurate vehicle control. To address this issue, we investigated the emergence of automatic driving rules using reinforcement learning based on information from surrounding vehicles using inter-vehicle communication. We evaluated whether reinforcement learning converges in a situation where distance sensor information can be shared in real-time using vehicle-to-vehicle communication and whether reinforcement learning can learn a rational driving method. The simulation results show a positive trend in the cumulative rewards value, and it indicates that the proposed multi-agent learning method with an extended own-vehicle environment has the potential to learn automated vehicle control with cooperative behavior automatically. Furthermore, we analyzed whether a rational driving method (action selection) can be learned by reinforcement learning. The simulation results showed that reinforcement learning achieves rational control of the overtaking behavior.
引用
收藏
页码:31 / 42
页数:12
相关论文
共 27 条
[1]   An Introduction to Deep Reinforcement Learning [J].
Francois-Lavet, Vincent ;
Henderson, Peter ;
Islam, Riashat ;
Bellemare, Marc G. ;
Pineau, Joelle .
FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2018, 11 (3-4) :219-354
[2]  
Haarnoja T, 2018, PR MACH LEARN RES, V80
[3]  
Haas J.K., 2014, Diss. Worcester Polytechnic Institute, V483, P484
[4]  
Hausknecht M., 2015, P AAAI FALL S SER, P29
[5]  
Juliani A, 2018, UNITY GEN PLATFORM I, DOI [10.48550/ARXIV.1809.02627, DOI 10.48550/ARXIV.1809.02627]
[6]   Study on the formulation of vehicle merging problems for model predictive control [J].
Kishi, Yuki ;
Cao, Wenjing ;
Mukai, Masakazu .
ARTIFICIAL LIFE AND ROBOTICS, 2022, 27 (03) :513-520
[7]   A Survey of Deep Learning Applications to Autonomous Vehicle Control [J].
Kuutti, Sampo ;
Bowden, Richard ;
Jin, Yaochu ;
Barber, Phil ;
Fallah, Saber .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (02) :712-733
[8]   A survey of 3D object detection algorithms for intelligent vehicles development [J].
Li, Zhen ;
Du, Yuren ;
Zhu, Miaomiao ;
Zhou, Shi ;
Zhang, Lifeng .
ARTIFICIAL LIFE AND ROBOTICS, 2022, 27 (01) :115-122
[9]   A Review of Shared Control for Automated Vehicles: Theory and Applications [J].
Marcano, Mauricio ;
Diaz, Sergio ;
Perez, Joshue ;
Irigoyen, Eloy .
IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2020, 50 (06) :475-491
[10]   Learning robust perceptive locomotion for quadrupedal robots in the wild [J].
Miki, Takahiro ;
Lee, Joonho ;
Hwangbo, Jemin ;
Wellhausen, Lorenz ;
Koltun, Vladlen ;
Hutter, Marco .
SCIENCE ROBOTICS, 2022, 7 (62)