Recent advances in hierarchical reinforcement learning

被引:3
作者
Barto, AG [1 ]
Mahadevan, S [1 ]
机构
[1] Univ Massachusetts, Dept Comp Sci, Autonomous Learning Lab, Amherst, MA 01003 USA
来源
DISCRETE EVENT DYNAMIC SYSTEMS-THEORY AND APPLICATIONS | 2003年 / 13卷 / 1-2期
基金
美国国家科学基金会;
关键词
D O I
10.1023/A:1022140919877
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning is bedeviled by the curse of dimensionality: the number of parameters to be learned grows exponentially with the size of any compact encoding of a state. Recent attempts to combat the curse of dimensionality have turned to principled ways of exploiting temporal abstraction, where decisions are not required at each step, but rather invoke the execution of temporally-extended activities which follow their own policies until termination. This leads naturally to hierarchical control architectures and associated learning algorithms. We review several approaches to temporal abstraction and hierarchical organization that machine learning researchers have recently developed. Common to these approaches is a reliance on the theory of semi-Markov decision processes, which we emphasize in our review. We then discuss extensions of these ideas to concurrent activities, multiagent coordination, and hierarchical memory for addressing partial observability. Concluding remarks address open challenges facing the further development of reinforcement learning in a hierarchical setting.
引用
收藏
页码:41 / 77
页数:37
相关论文
共 50 条
  • [41] Evaluating skills in hierarchical reinforcement learning
    Marzieh Davoodabadi Farahani
    Nasser Mozayani
    [J]. International Journal of Machine Learning and Cybernetics, 2020, 11 : 2407 - 2420
  • [42] Hierarchical Bayesian Inverse Reinforcement Learning
    Choi, Jaedeug
    Kim, Kee-Eung
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2015, 45 (04) : 793 - 805
  • [43] Hierarchical reinforcement learning and decision making
    Botvinick, Matthew Michael
    [J]. CURRENT OPINION IN NEUROBIOLOGY, 2012, 22 (06) : 956 - 962
  • [44] Modular Hierarchical Reinforcement Learning Algorithm
    Liu, Zhibin
    Zeng, Xiaoqin
    Liu, Huiyi
    [J]. INTELLIGENT COMPUTING THEORIES AND APPLICATIONS, ICIC 2012, 2012, 7390 : 375 - 382
  • [45] Evaluating skills in hierarchical reinforcement learning
    Farahani, Marzieh Davoodabadi
    Mozayani, Nasser
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (10) : 2407 - 2420
  • [46] Hierarchical Reinforcement Learning for Integrated Recommendation
    Xie, Ruobing
    Zhang, Shaoliang
    Wang, Rui
    Xia, Feng
    Lin, Leyu
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 4521 - 4528
  • [47] Autonomous Reinforcement Learning with Hierarchical REPS
    Daniel, Christian
    Neumann, Gerhard
    Peters, Jan
    [J]. 2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,
  • [48] Scalable Evolutionary Hierarchical Reinforcement Learning
    Abramowitz, Sasha
    Nitschke, Geoff
    [J]. PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 272 - 275
  • [49] A Neural Signature of Hierarchical Reinforcement Learning
    Ribas-Fernandes, Jose J. F.
    Solway, Alec
    Diuk, Carlos
    McGuire, Joseph T.
    Barto, Andrew G.
    Niv, Yael
    Botvinick, Matthew M.
    [J]. NEURON, 2011, 71 (02) : 370 - 379
  • [50] Hierarchical average reward reinforcement learning
    Department of Computing Science, University of Alberta, Edmonton, Alta. T6G 2E8, Canada
    不详
    [J]. Journal of Machine Learning Research, 2007, 8 : 2629 - 2669