Recent advances in hierarchical reinforcement learning

被引:3
作者
Barto, AG [1 ]
Mahadevan, S [1 ]
机构
[1] Univ Massachusetts, Dept Comp Sci, Autonomous Learning Lab, Amherst, MA 01003 USA
来源
DISCRETE EVENT DYNAMIC SYSTEMS-THEORY AND APPLICATIONS | 2003年 / 13卷 / 1-2期
基金
美国国家科学基金会;
关键词
D O I
10.1023/A:1022140919877
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Reinforcement learning is bedeviled by the curse of dimensionality: the number of parameters to be learned grows exponentially with the size of any compact encoding of a state. Recent attempts to combat the curse of dimensionality have turned to principled ways of exploiting temporal abstraction, where decisions are not required at each step, but rather invoke the execution of temporally-extended activities which follow their own policies until termination. This leads naturally to hierarchical control architectures and associated learning algorithms. We review several approaches to temporal abstraction and hierarchical organization that machine learning researchers have recently developed. Common to these approaches is a reliance on the theory of semi-Markov decision processes, which we emphasize in our review. We then discuss extensions of these ideas to concurrent activities, multiagent coordination, and hierarchical memory for addressing partial observability. Concluding remarks address open challenges facing the further development of reinforcement learning in a hierarchical setting.
引用
收藏
页码:41 / 77
页数:37
相关论文
共 50 条
  • [31] Hierarchical average reward reinforcement learning
    Ghavamzadeh, Mohammad
    Mahadevan, Sridhar
    JOURNAL OF MACHINE LEARNING RESEARCH, 2007, 8 : 2629 - 2669
  • [32] FeUdal Networks for Hierarchical Reinforcement Learning
    Vezhnevets, Alexander Sasha
    Osindero, Simon
    Schaul, Tom
    Heess, Nicolas
    Jaderberg, Max
    Silver, David
    Kavukcuoglu, Koray
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [33] A neural model of hierarchical reinforcement learning
    Rasmussen, Daniel
    Voelker, Aaron
    Eliasmith, Chris
    PLOS ONE, 2017, 12 (07):
  • [34] Hierarchical Reinforcement Learning for Quadruped Locomotion
    Jain, Deepali
    Iscen, Atil
    Caluwaerts, Ken
    2019 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2019, : 7551 - 7557
  • [35] Reinforcement Active Learning Hierarchical Loops
    Gordon, Goren
    Ahissar, Ehud
    2011 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2011, : 3008 - 3015
  • [36] Hierarchical Reinforcement Learning With Timed Subgoals
    Guertler, Nico
    Buechler, Dieter
    Martius, Georg
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [37] Reinforcement Learning From Hierarchical Critics
    Cao, Zehong
    Lin, Chin-Teng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) : 1066 - 1073
  • [38] Hierarchical Adversarial Inverse Reinforcement Learning
    Chen, Jiayu
    Lan, Tian
    Aggarwal, Vaneet
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17549 - 17558
  • [39] Partial Order Hierarchical Reinforcement Learning
    Hengst, Bernhard
    AI 2008: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2008, 5360 : 138 - 149
  • [40] Compositional Transfer in Hierarchical Reinforcement Learning
    Wulfmeier, Markus
    Abdolmaleki, Abbas
    Hafner, Roland
    Springenberg, Jost Tobias
    Neunert, Michael
    Hertweck, Tim
    Lampe, Thomas
    Siegel, Noah
    Heess, Nicolas
    Riedmiller, Martin
    ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,