Multi-Task Assisted Driving Policy Learning Method for Autonomous Driving

被引:0
|
作者
Luo, Yutao [1 ]
Xue, Zhicheng [1 ]
机构
[1] School of Mechanical and Automotive Engineering, Guangdong Provincial Key Laboratory of Automotive Engineering, South China University of Technology, Guangdong, Guangzhou
来源
Huanan Ligong Daxue Xuebao/Journal of South China University of Technology (Natural Science) | 2024年 / 52卷 / 10期
关键词
decision-making; driving policy; end-to-end autonomous driving; multi-task learning; reinforcement learning;
D O I
10.12141/j.issn.1000-565X.230503
中图分类号
学科分类号
摘要
With the development of autonomous driving technology, deep reinforcement learning has become an important means to realize the efficient driving policy learning. However, the implementation of autonomous driving is faced with the challenges brought by the complex and changeable traffic scenes, and the existing deep reinforcement learning methods have the problems of single scene adaptation ability and slow convergence speed. To address these issues and to improve the scene adaptability and policy learning efficiency of autonomous vehicles, this paper proposed a multi-task assisted driving policy learning method. Firstly, this method constructed the encoder-multi-task decoder module based on the deep residual network, squeezing high-dimensional driving scenes into low-dimensional representations, and adopted multi-task-assisted learning of semantic segmentation, depth estimation and speed prediction to improve the scene information richness of low-dimensional representations. Then, the low-dimensional representation was used as the state input to build a decision network based on reinforcement learning, and the multi-constraint reward function was designed to guide the learning of driving strategies. Finally, simulation experiments were conducted in CARLA. The experimental results show that, compared to classic methods such as DDPG and TD3, the proposed method improves the training process through multi-task assistance and learns better driving policies. It achieves higher task success rates and driving scores in several typical urban driving scenarios such as roundabouts and intersections, demonstrating excellent decision-making capabilities and scene adaptability. © 2024 South China University of Technology. All rights reserved.
引用
收藏
页码:31 / 40
页数:9
相关论文
共 23 条
  • [1] A comprehensive survey on the application of deep and reinforcement learning approaches in autonomous driving [J], Journal of King Saud University-Computer and Information Sciences, 34, 9, pp. 7366-7390, (2022)
  • [2] LIN Hongyi, Yang LIU, LI Shen, Research progress on key technologies in the cooperative vehicle infrastructure system, Journal of South China University of Technology(Natural Science Edition), 51, 10, pp. 46-67, (2023)
  • [3] Dynamic and interpretable state representation for deep reinforcement learning in automated driving [J], IFAC-PapersOnLine, 55, 24, pp. 129-134, (2022)
  • [4] HUANG C,, ZHANG R,, OUYANG M, Deductive reinforcement learning for visual autonomous urban driving navigation [J], IEEE Transactions on Neural Networks and Learning Systems, 32, 12, pp. 5379-5391, (2021)
  • [5] ZHU M, WANG Y., Human-like autonomous car-following model with deep reinforcement learning, Transportation Research Part C: Emerging Technologies, 97, pp. 348-368, (2018)
  • [6] KENDALL A, Learning to drive in a day [C] ∥ Proceedings of 2019 International Conference on Robotics and Automation, pp. 8248-8254, (2019)
  • [7] Proceedings of the 1st Conference on Robot Learning, pp. 1-16, (2017)
  • [8] SAXENA D M, Driving in dense traffic with model-free reinforcement learning, Proceedings of 2020 IEEE International Conference on Robotics and Automation, pp. 5385-5392, (2020)
  • [9] DEND Xiaohao, HOU Jin, TAN Guanghong, Multi-objective vehicle following decision algorithm based on reinforcement learning [J], Control and Decision, 36, 10, pp. 2497-2503, (2021)
  • [10] TOROMANOFF M,, WIRBEL E,, MOUTARDE F., End-to-end model-free reinforcement learning for urban driving using implicit affordances, Proceedings of 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7153-7162, (2020)