Fast Task Adaptation Based on the Combination of Model-Based and Gradient-Based Meta Learning

被引:12
|
作者
Xu, Zhixiong [1 ]
Chen, Xiliang [1 ]
Cao, Lei [1 ]
机构
[1] Army Engn Univ, Inst Command & Control Engn, Nanjing 210000, Peoples R China
关键词
Task analysis; Adaptation models; Reinforcement learning; Trajectory; Games; Data models; Training; Fast adaptation; gradient; metalearning; model-based; reinforcement learning;
D O I
10.1109/TCYB.2020.3028378
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep reinforcement learning (DRL) recently has attained remarkable results in various domains, including games, robotics, and recommender system. Nevertheless, an urgent problem in the practical application of DRL is fast adaptation. To this end, this article proposes a new and versatile metalearning approach called fast task adaptation via metalearning (FTAML), which leverages the strengths of the model-based methods and gradient-based metalearning methods for training the initial parameters of the model, such that the model is able to efficiently master unseen tasks with a little amount of data from the tasks. The proposed algorithm makes it possible to separate task optimization and task identification, specifically, the model-based learner helps to identify the pattern of a task, while the gradient-based metalearner is capable of consistently improving the performance with only a few gradient update steps through making use of the task embedding produced by the model-based learner. In addition, the choice of network for the model-based learner in the proposed method is also discussed, and the performance of networks with different depths is explored. Finally, the simulation results on reinforcement learning problems demonstrate that the proposed approach outperforms compared metalearning algorithms and delivers a new state-of-the-art performance on a variety of challenging control tasks.
引用
收藏
页码:5209 / 5218
页数:10
相关论文
共 50 条
  • [31] Model-based learning protects against forming habits
    Gillan, Claire M.
    Otto, A. Ross
    Phelps, Elizabeth A.
    Daw, Nathaniel D.
    COGNITIVE AFFECTIVE & BEHAVIORAL NEUROSCIENCE, 2015, 15 (03) : 523 - 536
  • [32] Model-Based OPC With Adaptive PID Control Through Reinforcement Learning
    Kim, Taeyoung
    Zhang, Shilong
    Shin, Youngsoo
    IEEE TRANSACTIONS ON SEMICONDUCTOR MANUFACTURING, 2025, 38 (01) : 48 - 56
  • [33] Fast Human-in-the-Loop Control for HVAC Systems via Meta-Learning and Model-Based Offline Reinforcement Learning
    Chen, Liangliang
    Meng, Fei
    Zhang, Ying
    IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING, 2023, 8 (03): : 504 - 521
  • [34] A Model-Based GNN for Learning Precoding
    Guo, Jia
    Yang, Chenyang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (07) : 6983 - 6999
  • [35] A survey on model-based reinforcement learning
    Luo, Fan-Ming
    Xu, Tian
    Lai, Hang
    Chen, Xiong-Hui
    Zhang, Weinan
    Yu, Yang
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (02)
  • [36] Model-based learning protects against forming habits
    Claire M. Gillan
    A. Ross Otto
    Elizabeth A. Phelps
    Nathaniel D. Daw
    Cognitive, Affective, & Behavioral Neuroscience, 2015, 15 : 523 - 536
  • [37] Learnable Weighting Mechanism in Model-based Reinforcement Learning
    Huang W.-Z.
    Yin Q.-Y.
    Zhang J.-G.
    Huang K.-Q.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (06): : 2765 - 2775
  • [38] Personalized task difficulty adaptation based on reinforcement learning
    Yaqian Zhang
    Wooi-Boon Goh
    User Modeling and User-Adapted Interaction, 2021, 31 : 753 - 784
  • [39] Gradient-Based Fast Intra Coding Decision Algorithm for HEVC
    Wang, Yuting
    Cao, Jian
    Wang, Jun
    Liang, Fan
    2019 IEEE 4TH INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING (ICSIP 2019), 2019, : 870 - 874
  • [40] Personalized task difficulty adaptation based on reinforcement learning
    Zhang, Yaqian
    Goh, Wooi-Boon
    USER MODELING AND USER-ADAPTED INTERACTION, 2021, 31 (04) : 753 - 784