Spatiotemporal Costmap Inference for MPC Via Deep Inverse Reinforcement Learning

被引:15
作者
Lee, Keuntaek [1 ]
Isele, David [2 ]
Theodorou, Evangelos A. [3 ]
Bae, Sangjae [2 ]
机构
[1] Georgia Inst Technol, Dept Elect & Comp Engn, Atlanta, GA 30318 USA
[2] Honda Res Inst USA Inc, Div Res, San Jose, CA 95110 USA
[3] Georgia Inst Technol, Sch Aerosp Engn, Atlanta, GA 30318 USA
关键词
Learning from demonstration; reinforcement learning; optimization and optimal control; motion and path planning; autonomous vehicle navigation;
D O I
10.1109/LRA.2022.3146635
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
It can he difficult to autonomously produce driver behavior so that it appears natural to other traffic participants. Through Inverse Reinforcement Learning (IRL), we can automate this process by learning the underlying reward function from human demonstrations. We propose a new IRL algorithm that learns a goal-conditioned spatio-temporal reward function. The resulting costmap is used by Model Predictive Controllers (MPCs) to perform a task without any hand-designing or hand-tuning of the cost function. We evaluate our proposed Goal-conditioned SpatioTemporal Zeroing Maximum Entropy Deep IRL (GSTZ)-MEDIRL framework together with MPC in the CARLA simulator for autonomous driving, lane keeping, and lane changing tasks in a challenging dense traffic highway scenario. Our proposed methods show higher success rates compared to other baseline methods including behavior cloning, state-of-the-art RL policies, and MPC with a learning-based behavior prediction model.
引用
收藏
页码:3194 / 3201
页数:8
相关论文
共 50 条
  • [41] Empirical study of privacy inference attack against deep reinforcement learning models
    Zhou, Huaicheng
    Mo, Kanghua
    Huang, Teng
    Li, Yongjin
    CONNECTION SCIENCE, 2023, 35 (01)
  • [42] Membership Inference Attacks Against Temporally Correlated Data in Deep Reinforcement Learning
    Gomrokchi, Maziar
    Amin, Susan
    Aboutalebi, Hossein
    Wong, Alexander
    Precup, Doina
    IEEE ACCESS, 2023, 11 : 42796 - 42808
  • [43] The Advance of Reinforcement Learning and Deep Reinforcement Learning
    Lyu, Le
    Shen, Yang
    Zhang, Sicheng
    2022 IEEE INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, BIG DATA AND ALGORITHMS (EEBDA), 2022, : 644 - 648
  • [44] Fvading Deep Learning -Based Malware Detectors via Obfuscation: A Deep Reinforcement Learning Approach
    Etter, Brian
    Hu, James Lee
    Ebrahimi, Mohammadreza
    Li, Weifeng
    Li, Xin
    Chen, Hsinchun
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 101 - 109
  • [45] A novel inverse design method for morphing airfoil based on deep reinforcement learning
    Su, Jing
    Sun, Gang
    Tao, Jun
    AEROSPACE SCIENCE AND TECHNOLOGY, 2024, 145
  • [46] Robo-Advising: Enhancing Investment with Inverse Optimization and Deep Reinforcement Learning
    Wang, Haoran
    Yu, Shi
    20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 365 - 372
  • [47] Market Making Strategy Optimization via Deep Reinforcement Learning
    Sun, Tianyuan
    Huang, Dechun
    Yu, Jie
    IEEE ACCESS, 2022, 10 : 9085 - 9093
  • [48] Dynamical Hyperparameter Optimization via Deep Reinforcement Learning in Tracking
    Dong, Xingping
    Shen, Jianbing
    Wang, Wenguan
    Shao, Ling
    Ling, Haibin
    Porikli, Fatih
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (05) : 1515 - 1529
  • [49] Discriminative sampling via deep reinforcement learning for kinship verification
    Wang, Shiwei
    Yan, Haibin
    PATTERN RECOGNITION LETTERS, 2020, 138 : 38 - 43
  • [50] Solving the train dispatching problem via deep reinforcement learning
    Agasucci, Valerio
    Grani, Giorgio
    Lamorgese, Leonardo
    JOURNAL OF RAIL TRANSPORT PLANNING & MANAGEMENT, 2023, 26