Deep Surrogate Q-Learning for Autonomous Driving

被引:3
|
作者
Kalweit, Maria [1 ]
Kalweit, Gabriel [1 ]
Werling, Moritz [2 ]
Boedecker, Joschka [1 ,3 ]
机构
[1] Univ Freiburg, Neuroroboties Lab, Freiburg, Germany
[2] BMWGroup, Unterschleissheim, Germany
[3] Univ Freiburg, BrainLinks BrainTools, Freiburg, Germany
关键词
D O I
10.1109/ICRA46639.2022.9811618
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Open challenges for deep reinforcement learning systems are their adaptivity to changing environments and their efficiency w.r.t. computational resources and data. In the application of learning lane-change behavior for autonomous driving, the number of required transitions imposes a bottleneck, since test drivers cannot perform an arbitrary amount of lane changes in the real world. In the off-policy setting, additional information on solving the task can be gained by observing actions from others. While in the classical RL setup this knowledge remains unused, we use other drivers as surrogates to learn the agent's value function more efficiently. We propose Surrogate Q-learning that deals with the aforementioned problems and reduces the required driving time drastically. We further propose an efficient implementation based on a permutation equivariant deep neural network architecture of the Q-function to estimate action-values for a variable number of vehicles in sensor range. We evaluate our method in the open traffic simulator SUMO and learn well performing driving policies on the real highD dataset.
引用
收藏
页码:1578 / 1584
页数:7
相关论文
共 50 条
  • [1] Motion Primitives as the Action Space of Deep Q-Learning for Planning in Autonomous Driving
    Schneider, Tristan
    Pedrosa, Matheus V. A.
    Gros, Timo P.
    Wolf, Verena
    Flasskamp, Kathrin
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (11) : 17852 - 17864
  • [2] Autonomous Driving in Roundabout Maneuvers Using Reinforcement Learning with Q-Learning
    Garcia Cuenca, Laura
    Puertas, Enrique
    Fernandez Andres, Javier
    Aliane, Nourdine
    ELECTRONICS, 2019, 8 (12)
  • [3] Autonomous Warehouse Robot using Deep Q-Learning
    Peyas, Ismot Sadik
    Hasan, Zahid
    Tushar, Md Rafat Rahman
    Al Musabbir
    Azni, Raisa Mehjabin
    Siddique, Shahnewaz
    2021 IEEE REGION 10 CONFERENCE (TENCON 2021), 2021, : 857 - 862
  • [4] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483
  • [5] DQ-GAT: Towards Safe and Efficient Autonomous Driving With Deep Q-Learning and Graph Attention Networks
    Cai, Peide
    Wang, Hengli
    Sun, Yuxiang
    Liu, Ming
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (11) : 21102 - 21112
  • [6] Recurrent Deep Multiagent Q-Learning for Autonomous Brokers in Smart Grid
    Yang, Yaodong
    Hao, Jianye
    Sun, Mingyang
    Wang, Zan
    Fan, Changjie
    Strbac, Goran
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 569 - 575
  • [7] Q-Learning for autonomous vehicle navigation
    Gonzalez-Miranda, Oscar
    Miranda, Luis Antonio Lopez
    Ibarra-Zannatha, Juan Manuel
    2023 XXV ROBOTICS MEXICAN CONGRESS, COMROB, 2023, : 138 - 142
  • [8] Constrained Deep Q-Learning Gradually Approaching Ordinary Q-Learning
    Ohnishi, Shota
    Uchibe, Eiji
    Yamaguchi, Yotaro
    Nakanishi, Kosuke
    Yasui, Yuji
    Ishii, Shin
    FRONTIERS IN NEUROROBOTICS, 2019, 13
  • [9] Optimizing Agent Training with Deep Q-Learning on a Self Driving Reinforcement Learning Environment
    Rodrigues, Pedro
    Vieira, Susana
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 745 - 752
  • [10] Amortized Q-learning with Model-based Action Proposals for Autonomous Driving on Highways
    Mirchevska, Branka
    Huegle, Maria
    Kalweit, Gabriel
    Werling, Moritz
    Boedecker, Joschka
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 1028 - 1035