Inverse reinforcement learning from summary data

被引:0
|
作者
Antti Kangasrääsiö
Samuel Kaski
机构
[1] Aalto University,Department of Computer Science
来源
Machine Learning | 2018年 / 107卷
关键词
Inverse reinforcement learning; Bayesian inference; Monte-Carlo estimation; Approximate Bayesian computation;
D O I
暂无
中图分类号
学科分类号
摘要
Inverse reinforcement learning (IRL) aims to explain observed strategic behavior by fitting reinforcement learning models to behavioral data. However, traditional IRL methods are only applicable when the observations are in the form of state-action paths. This assumption may not hold in many real-world modeling settings, where only partial or summarized observations are available. In general, we may assume that there is a summarizing function σ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sigma $$\end{document}, which acts as a filter between us and the true state-action paths that constitute the demonstration. Some initial approaches to extending IRL to such situations have been presented, but with very specific assumptions about the structure of σ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sigma $$\end{document}, such as that only certain state observations are missing. This paper instead focuses on the most general case of the problem, where no assumptions are made about the summarizing function, except that it can be evaluated. We demonstrate that inference is still possible. The paper presents exact and approximate inference algorithms that allow full posterior inference, which is particularly important for assessing parameter uncertainty in this challenging inference situation. Empirical scalability is demonstrated to reasonably sized problems, and practical applicability is demonstrated by estimating the posterior for a cognitive science RL model based on an observed user’s task completion time only.
引用
收藏
页码:1517 / 1535
页数:18
相关论文
共 50 条
  • [31] Identifiability and Generalizability from Multiple Experts in Inverse Reinforcement Learning
    Rolland, Paul
    Viano, Luca
    Schurhoff, Norman
    Nikolov, Boris
    Cevher, Volkan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [32] Inverse Reinforcement Learning from a Gradient-based Learner
    Ramponi, Giorgia
    Drappo, Gianluca
    Restelli, Marcello
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [33] Reinforcement Learning and Inverse Reinforcement Learning with System 1 and System 2
    Peysakhovich, Alexander
    AIES '19: PROCEEDINGS OF THE 2019 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2019, : 409 - 415
  • [34] Approximate Inverse Reinforcement Learning from Vision-based Imitation Learning
    Lee, Keuntaek
    Vlahov, Bogdan
    Gibson, Jason
    Rehg, James M.
    Theodorou, Evangelos A.
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 10793 - 10799
  • [35] Data-Driven Inverse Reinforcement Learning Control for Linear Multiplayer Games
    Lian, Bosen
    Donge, Vrushabh S.
    Lewis, Frank L.
    Chai, Tianyou
    Davoudi, Ali
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (02) : 2028 - 2041
  • [36] Certified data-driven inverse reinforcement learning of Markov jump systems
    Xue, Wenqian
    Lewis, Frank L.
    Lian, Bosen
    AUTOMATICA, 2025, 176
  • [37] Inverse Reinforcement Learning for Text Summarization
    Fu, Yu
    Xiong, Deyi
    Dong, Yue
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 6559 - 6570
  • [38] Data Driven Strategies for Active Monocular SLAM using Inverse Reinforcement Learning
    Prasad, Vignesh
    Jangir, Rishabh
    Balaraman, Ravindran
    Krishna, K. Madhava
    AAMAS'17: PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2017, : 1697 - 1699
  • [39] Reward Identification in Inverse Reinforcement Learning
    Kim, Kuno
    Garg, Shivam
    Shiragur, Kirankumar
    Ermon, Stefano
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [40] Compatible Reward Inverse Reinforcement Learning
    Metelli, Alberto Maria
    Pirotta, Matteo
    Restelli, Marcello
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30