MetaDrive: Composing Diverse Driving Scenarios for Generalizable Reinforcement Learning

被引:48
|
作者
Li, Quanyi [1 ]
Peng, Zhenghao [2 ]
Feng, Lan [4 ]
Zhang, Qihang [3 ]
Xue, Zhenghai [3 ]
Zhou, Bolei [5 ]
机构
[1] Chinese Univ Hong Kong, Ctr Perceptual & Interact Intelligence, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[3] Chinese Univ Hong Kong, Dept Informat Engn, Hong Kong, Peoples R China
[4] Swiss Fed Inst Technol, CH-8092 Zurich, Switzerland
[5] Univ Calif Los Angeles, Los Angeles, CA 90095 USA
关键词
Task analysis; Roads; Reinforcement learning; Benchmark testing; Training; Safety; Autonomous vehicles; autonomous driving; simulation;
D O I
10.1109/TPAMI.2022.3190471
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Driving safely requires multiple capabilities from human and intelligent agents, such as the generalizability to unseen environments, the safety awareness of the surrounding traffic, and the decision-making in complex multi-agent settings. Despite the great success of Reinforcement Learning (RL), most of the RL research works investigate each capability separately due to the lack of integrated environments. In this work, we develop a new driving simulation platform called MetaDrive to support the research of generalizable reinforcement learning algorithms for machine autonomy. MetaDrive is highly compositional, which can generate an infinite number of diverse driving scenarios from both the procedural generation and the real data importing. Based on MetaDrive, we construct a variety of RL tasks and baselines in both single-agent and multi-agent settings, including benchmarking generalizability across unseen scenes, safe exploration, and learning multi-agent traffic. The generalization experiments conducted on both procedurally generated scenarios and real-world scenarios show that increasing the diversity and the size of the training set leads to the improvement of the RL agent's generalizability. We further evaluate various safe reinforcement learning and multi-agent reinforcement learning algorithms in MetaDrive environments and provide the benchmarks. Source code, documentation, and demo video are available at https://metadriverse.github.io/metadrive.
引用
收藏
页码:3461 / 3475
页数:15
相关论文
共 50 条
  • [41] Deep Reinforcement Learning for Autonomous Driving in Amazon Web Services DeepRacer
    Petryshyn, Bohdan
    Postupaiev, Serhii
    Ben Bari, Soufiane
    Ostreika, Armantas
    INFORMATION, 2024, 15 (02)
  • [42] Deep Reinforcement Learning on Autonomous Driving Policy With Auxiliary Critic Network
    Wu, Yuanqing
    Liao, Siqin
    Liu, Xiang
    Li, Zhihang
    Lu, Renquan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (07) : 3680 - 3690
  • [43] Generalizable Reinforcement Learning-Based Coarsening Model for Resource Allocation over Large and Diverse Stream Processing Graphs
    Nie, Lanshun
    Qiu, Yuqi
    Meng, Fei
    Yu, Mo
    Li, Jing
    2023 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, IPDPS, 2023, : 435 - 445
  • [44] Decision-Making for Autonomous Vehicles in Random Task Scenarios at Unsignalized Intersection Using Deep Reinforcement Learning
    Xiao, Wenxuan
    Yang, Yuyou
    Mu, Xinyu
    Xie, Yi
    Tang, Xiaolin
    Cao, Dongpu
    Liu, Teng
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (06) : 7812 - 7825
  • [45] Interaction-Aware Planning With Deep Inverse Reinforcement Learning for Human-Like Autonomous Driving in Merge Scenarios
    Nan, Jiangfeng
    Deng, Weiwen
    Zhang, Ruzheng
    Wang, Ying
    Zhao, Rui
    Ding, Juan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 2714 - 2726
  • [46] Automatically Learning Fallback Strategies with Model-Free Reinforcement Learning in Safety-Critical Driving Scenarios
    Lecerf, Ugo U. L.
    Yemdji-Tchassi, Christelle C. Y.
    Aubert, Sebastien S. A.
    Michiardi, Pietro P. M.
    PROCEEDINGS OF 2022 7TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING TECHNOLOGIES, ICMLT 2022, 2022, : 209 - 215
  • [47] Generalizable Crowd Counting via Diverse Context Style Learning
    Zhao, Wenda
    Wang, Mingyue
    Liu, Yu
    Lu, Huimin
    Xu, Congan
    Yao, Libo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5399 - 5410
  • [48] Reinforcement Learning for Shared Driving
    Ko, Sangjin
    Langari, Reza
    IFAC PAPERSONLINE, 2023, 56 (03): : 247 - 252
  • [49] A Task-Agnostic Regularizer for Diverse Subpolicy Discovery in Hierarchical Reinforcement Learning
    Huo, Liangyu
    Wang, Zulin
    Xu, Mai
    Song, Yuhang
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (03): : 1932 - 1944
  • [50] Deep Reinforcement Learning With NMPC Assistance Nash Switching for Urban Autonomous Driving
    Alighanbari, Sina
    Azad, Nasser L.
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (03): : 2604 - 2615