Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning

被引:15
作者
Lee, Keuntaek [1 ]
Isele, David [2 ]
Theodorou, Evangelos A. [3 ]
Bae, Sangjae [2 ]
机构
[1] Georgia Inst Technol, Dept Elect & Comp Engn, Atlanta, GA 30318 USA
[2] Honda Res Inst USA Inc, Div Res, San Jose, CA 95110 USA
[3] Georgia Inst Technol, Sch Aerosp Engn, Atlanta, GA 30318 USA
关键词
Learning from demonstration; reinforcement learning; optimization and optimal control; motion and path planning; autonomous vehicle navigation;
D O I
10.1109/LRA.2022.3146635
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
It can he difficult to autonomously produce driver behavior so that it appears natural to other traffic participants. Through Inverse Reinforcement Learning (IRL), we can automate this process by learning the underlying reward function from human demonstrations. We propose a new IRL algorithm that learns a goal-conditioned spatio-temporal reward function. The resulting costmap is used by Model Predictive Controllers (MPCs) to perform a task without any hand-designing or hand-tuning of the cost function. We evaluate our proposed Goal-conditioned SpatioTemporal Zeroing Maximum Entropy Deep IRL (GSTZ)-MEDIRL framework together with MPC in the CARLA simulator for autonomous driving, lane keeping, and lane changing tasks in a challenging dense traffic highway scenario. Our proposed methods show higher success rates compared to other baseline methods including behavior cloning, state-of-the-art RL policies, and MPC with a learning-based behavior prediction model.
引用
收藏
页码:3194 / 3201
页数:8
相关论文
共 50 条
  • [21] Data distribution inference attack in federated learning via reinforcement learning support
    Yu, Dongxiao
    Zhang, Hengming
    Huang, Yan
    Xie, Zhenzhen
    HIGH-CONFIDENCE COMPUTING, 2025, 5 (01):
  • [22] Network-Scale Traffic Signal Control via Multiagent Reinforcement Learning With Deep Spatiotemporal Attentive Network
    Huang, Hao
    Hu, Zhiqun
    Lu, Zhaoming
    Wen, Xiangming
    IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (01) : 262 - 274
  • [23] Learning to Navigate in Human Environments via Deep Reinforcement Learning
    Gao, Xingyuan
    Sun, Shiying
    Zhao, Xiaoguang
    Tan, Min
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 418 - 429
  • [24] A CGRA based Neural Network Inference Engine for Deep Reinforcement Learning
    Liang, Minglan
    Chen, Mingsong
    Wang, Zheng
    Sun, Jingwei
    2018 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2018), 2018, : 540 - 543
  • [25] A Novel Framework Combining MPC and Deep Reinforcement Learning With Application to Freeway Traffic Control
    Sun, Dingshan
    Jamshidnejad, Anahita
    De Schutter, Bart
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (07) : 6756 - 6769
  • [26] Traffic Signal Timing via Deep Reinforcement Learning
    Li, Li
    Lv, Yisheng
    Wang, Fei-Yue
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2016, 3 (03) : 247 - 254
  • [27] Underwater Color Correction via Deep Reinforcement Learning
    Sun, Shixin
    Wang, Hao
    Wu, Xiaohui
    Li, Li
    Zhang, Hao
    Li, Mingjie
    Ren, Peng
    OCEANS 2021: SAN DIEGO - PORTO, 2021,
  • [28] Advanced Building Control via Deep Reinforcement Learning
    Jia, Ruoxi
    Jin, Ming
    Sun, Kaiyu
    Hong, Tianzhen
    Spanos, Costas
    INNOVATIVE SOLUTIONS FOR ENERGY TRANSITIONS, 2019, 158 : 6158 - 6163
  • [29] Deep Inverse Reinforcement Learning for Objective Function Identification in Bidding Models
    Guo, Hongye
    Chen, Qixin
    Xia, Qing
    Kang, Chongqing
    IEEE TRANSACTIONS ON POWER SYSTEMS, 2021, 36 (06) : 5684 - 5696
  • [30] Road Planning for Slums via Deep Reinforcement Learning
    Zheng, Yu
    Su, Hongyuan
    Ding, Jingtao
    Jin, Depeng
    Li, Yong
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 5695 - 5706