A Lower Bound for the Sample Complexity of Inverse Reinforcement Learning

被引:0
|
作者
Komanduru, Abi [1 ]
Honorio, Jean [1 ]
机构
[1] Purdue Univ, W Lafayette, IN 47907 USA
来源
INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139 | 2021年 / 139卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Inverse reinforcement learning (IRL) is the task of finding a reward function that generates a desired optimal policy for a given Markov Decision Process (MDP). This paper develops an informationtheoretic lower bound for the sample complexity of the finite state, finite action IRL problem. A geometric construction of beta-strict separable IRL problems using spherical codes is considered. Properties of the ensemble size as well as the Kullback-Leibler divergence between the generated trajectories are derived. The resulting ensemble is then used along with Fano's inequality to derive a sample complexity lower bound of O(n log n), where n is the number of states in the MDP.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] On the Correctness and Sample Complexity of Inverse Reinforcement Learning
    Komanduru, Abi
    Honorio, Jean
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Balancing Sample Efficiency and Suboptimality in Inverse Reinforcement Learning
    Damiani, Angelo
    Manganini, Giorgio
    Metelli, Alberto Maria
    Restelli, Marcello
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [3] Sample Complexity of Robust Reinforcement Learning with a Generative Model
    Panaganti, Kishan
    Kalathil, Dileep
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [4] Learning with Safety Constraints: Sample Complexity of Reinforcement Learning for Constrained MDPs
    HasanzadeZonuzy, Aria
    Bura, Archana
    Kalathil, Dileep
    Shakkottai, Srinivas
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7667 - 7674
  • [5] An Improved Sample Complexity Lower Bound for (Fidelity) Quantum State Tomography
    Yuen, Henry
    QUANTUM, 2023, 7 : 1 - 4
  • [6] Sample Complexity of Model-Based Robust Reinforcement Learning
    Panaganti, Kishan
    Kalathil, Dileep
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 2240 - 2245
  • [7] Sample Complexity of Goal-Conditioned Hierarchical Reinforcement Learning
    Robert, Arnaud
    Pike-Burke, Ciara
    Faisal, A. Aldo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [8] The Sample Complexity of Teaching-by-Reinforcement on Q-Learning
    Zhang, Xuezhou
    Bharti, Shubham Kumar
    Ma, Yuzhe
    Singla, Adish
    Zhu, Xiaojin
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10939 - 10947
  • [9] Sample Complexity of Episodic Fixed-Horizon Reinforcement Learning
    Dann, Christoph
    Brunskill, Emma
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [10] Settling the Horizon-Dependence of Sample Complexity in Reinforcement Learning
    Li, Yuanzhi
    Wang, Ruosong
    Yang, Lin F.
    2021 IEEE 62ND ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS 2021), 2022, : 965 - 976