Discovering and Exploiting Skills in Hierarchical Reinforcement Learning

被引:0
|
作者
Huang, Zhigang [1 ]
机构
[1] China Ship Sci Res Ctr, Wuxi 214062, Peoples R China
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Reinforcement learning; Trajectory; Turning; Training; Time series analysis; Switches; Planning; Entropy; Wheels; Recurrent neural networks; Hierarchical reinforcement learning; skill discovery; skill exploitation; exploration;
D O I
10.1109/ACCESS.2024.3491339
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Humans can perform infinite diverse skills. These skills typically represent abstract knowledge that is highly correlated with time series. To behave more like a human, we take a long-term planning perspective to discover and exploit skills (DES) in hierarchical reinforcement learning. We view the skill-learning process as an extension of primitive skills to advanced skills and ensure that they have sufficient exploration capability. DES discovers skills at the level of a trajectory sequence within a skill length, rather than at the level of individual states and actions. It assigns the skill inference loss from the recurrent neural network evenly to each time step, maximizing skill differentiation to cover fine-grained local areas. Furthermore, DES exploits skills in an adaptive way. It builds on a multi-step combination, and then makes switching decisions according to the relative advantages of the previous and the estimated skills, thus achieving long-term form skills. These advanced skills allow the agent to escape from local areas without sacrificing flexibility. A skill truncation is also set to prevent excessive exploration. Moreover, we verify the necessity of our discovery and exploitation methods from the perspective of skill inference and exploration capability, respectively. Our experimental analysis demonstrates the superiority of DES on continuous control tasks with sparse rewards and explains the benefits of our methods.
引用
收藏
页码:163042 / 163055
页数:14
相关论文
共 50 条
  • [1] Evaluating skills in hierarchical reinforcement learning
    Farahani, Marzieh Davoodabadi
    Mozayani, Nasser
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (10) : 2407 - 2420
  • [2] Evaluating skills in hierarchical reinforcement learning
    Marzieh Davoodabadi Farahani
    Nasser Mozayani
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 2407 - 2420
  • [3] Skill-Critic: Refining Learned Skills for Hierarchical Reinforcement Learning
    Hao, Ce
    Weaver, Catherine
    Tang, Chen
    Kawamoto, Kenta
    Tomizuka, Masayoshi
    Zhan, Wei
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (04): : 3625 - 3632
  • [4] Feature Control as Intrinsic Motivation for Hierarchical Reinforcement Learning
    Dilokthanakul, Nat
    Kaplanis, Christos
    Pawlowski, Nick
    Shanahan, Murray
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (11) : 3409 - 3418
  • [5] Learning disentangled skills for hierarchical reinforcement learning through trajectory autoencoder with weak labels
    Song, Wonil
    Jeon, Sangryul
    Choi, Hyesong
    Sohn, Kwanghoon
    Min, Dongbo
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 230
  • [6] Discovering Intrinsic Subgoals for Vision-and-Language Navigation via Hierarchical Reinforcement Learning
    Wang, Jiawei
    Wang, Teng
    Xu, Lele
    He, Zichen
    Sun, Changyin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (04) : 6516 - 6528
  • [7] MENTOR: Guiding Hierarchical Reinforcement Learning With Human Feedback and Dynamic Distance Constraint
    Zhou, Xinglin
    Yuan, Yifu
    Yang, Shaofu
    Hao, Jianye
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025,
  • [8] Hierarchical reinforcement learning with subpolicies specializing for learned subgoals
    Bakker, B
    Schmidhuber, J
    Proceedings of the Second IASTED International Conference on Neural Networks and Computational Intelligence, 2004, : 125 - 130
  • [9] Developing Driving Strategies Efficiently: A Skill-Based Hierarchical Reinforcement Learning Approach
    Gurses, Yigit
    Buyukdemirci, Kaan
    Yildiz, Yildiray
    IEEE CONTROL SYSTEMS LETTERS, 2024, 8 : 121 - 126
  • [10] Reinforcement Learning From Hierarchical Critics
    Cao, Zehong
    Lin, Chin-Teng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (02) : 1066 - 1073