Deep Surrogate Q-Learning for Autonomous Driving

被引:3
|
作者
Kalweit, Maria [1 ]
Kalweit, Gabriel [1 ]
Werling, Moritz [2 ]
Boedecker, Joschka [1 ,3 ]
机构
[1] Univ Freiburg, Neuroroboties Lab, Freiburg, Germany
[2] BMWGroup, Unterschleissheim, Germany
[3] Univ Freiburg, BrainLinks BrainTools, Freiburg, Germany
关键词
D O I
10.1109/ICRA46639.2022.9811618
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Open challenges for deep reinforcement learning systems are their adaptivity to changing environments and their efficiency w.r.t. computational resources and data. In the application of learning lane-change behavior for autonomous driving, the number of required transitions imposes a bottleneck, since test drivers cannot perform an arbitrary amount of lane changes in the real world. In the off-policy setting, additional information on solving the task can be gained by observing actions from others. While in the classical RL setup this knowledge remains unused, we use other drivers as surrogates to learn the agent's value function more efficiently. We propose Surrogate Q-learning that deals with the aforementioned problems and reduces the required driving time drastically. We further propose an efficient implementation based on a permutation equivariant deep neural network architecture of the Q-function to estimate action-values for a variable number of vehicles in sensor range. We evaluate our method in the open traffic simulator SUMO and learn well performing driving policies on the real highD dataset.
引用
收藏
页码:1578 / 1584
页数:7
相关论文
共 50 条
  • [11] A Modular Autonomous Driving System for Electric Boats based on Fuzzy Controllers and Q-Learning
    Ferrandino, Emanuele
    Capillo, Antonino
    De Santis, Enrico
    Mascioli, Fabio M. F.
    Rizzi, Antonello
    PROCEEDINGS OF THE 13TH INTERNATIONAL JOINT CONFERENCE ON COMPUTATIONAL INTELLIGENCE (IJCCI), 2021, : 185 - 195
  • [12] Recurrent Deep Multiagent Q-Learning for Autonomous Agents in Future Smart Grid
    Yang, Yaodong
    Hao, Jianye
    Wang, Zan
    Sun, Mingyang
    Strbac, Goran
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 2136 - 2138
  • [13] Detection of Static and Mobile Targets by an Autonomous Agent with Deep Q-Learning Abilities
    Matzliach, Barouch
    Ben-Gal, Irad
    Kagan, Evgeny
    ENTROPY, 2022, 24 (08)
  • [14] Deep Q Learning for Traffic Simulation in Autonomous Driving at a Highway Junction
    Kashihara, Koji
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 984 - 988
  • [15] Deep Learning for Autonomous Driving
    Burleigh, Nicholas
    King, Jordan
    Braunl, Thomas
    2019 DIGITAL IMAGE COMPUTING: TECHNIQUES AND APPLICATIONS (DICTA), 2019, : 105 - 112
  • [16] Interactive Q-learning to adapt autonomous interface
    Ishiwaka, Y
    Takehara, N
    Yokoi, H
    Kakazu, Y
    INTELLIGENT AUTONOMOUS VEHICLES 2001, 2002, : 45 - 50
  • [17] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100
  • [18] Comparison of Deep Q-Learning, Q-Learning and SARSA Reinforced Learning for Robot Local Navigation
    Anas, Hafiq
    Ong, Wee Hong
    Malik, Owais Ahmed
    ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS 6, 2022, 429 : 443 - 454
  • [19] Hierarchical clustering with deep Q-learning
    Forster, Richard
    Fulop, Agnes
    ACTA UNIVERSITATIS SAPIENTIAE INFORMATICA, 2018, 10 (01) : 86 - 109
  • [20] Active deep Q-learning with demonstration
    Si-An Chen
    Voot Tangkaratt
    Hsuan-Tien Lin
    Masashi Sugiyama
    Machine Learning, 2020, 109 : 1699 - 1725