Pre-training with Augmentations for Efficient Transfer in Model-Based Reinforcement Learning

被引:0
|
作者
Esteves, Bernardo [1 ,2 ]
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
机构
[1] INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
来源
PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2023, PT I | 2023年 / 14115卷
关键词
Reinforcement learning; Transfer learning; Representation learning;
D O I
10.1007/978-3-031-49008-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work explores pre-training as a strategy to allow reinforcement learning (RL) algorithms to efficiently adapt to new (albeit similar) tasks. We argue for introducing variability during the pre-training phase, in the form of augmentations to the observations of the agent, to improve the sample efficiency of the fine-tuning stage. We categorize such variability in the form of perceptual, dynamic and semantic augmentations, which can be easily employed in standard pre-training methods. We perform extensive evaluations of our proposed augmentation scheme in model-based algorithms, across multiple scenarios of increasing complexity. The results consistently show that our augmentation scheme significantly improves the efficiency of the fine-tuning to novel tasks, outperforming other state-of-the-art pre-training approaches.
引用
收藏
页码:133 / 145
页数:13
相关论文
共 50 条
  • [21] Efficient hyperparameters optimization through model-based reinforcement learning with experience exploiting and meta-learning
    Liu, Xiyuan
    Wu, Jia
    Chen, Senpeng
    SOFT COMPUTING, 2023, 27 (13) : 8661 - 8678
  • [22] Efficient hyperparameters optimization through model-based reinforcement learning with experience exploiting and meta-learning
    Xiyuan Liu
    Jia Wu
    Senpeng Chen
    Soft Computing, 2023, 27 : 8661 - 8678
  • [23] Reinforcement Twinning: From digital twins to model-based reinforcement learning
    Schena, Lorenzo
    Marques, Pedro A.
    Poletti, Romain
    Van den Berghe, Jan
    Mendez, Miguel A.
    JOURNAL OF COMPUTATIONAL SCIENCE, 2024, 82
  • [24] Latent Causal Dynamics Model for Model-Based Reinforcement Learning
    Hao, Zhifeng
    Zhu, Haipeng
    Chen, Wei
    Cai, Ruichu
    NEURAL INFORMATION PROCESSING, ICONIP 2023, PT II, 2024, 14448 : 219 - 230
  • [25] Model-based reinforcement learning with model error and its application
    Tajima, Yoshiyuki
    Onisawa, Takehisa
    PROCEEDINGS OF SICE ANNUAL CONFERENCE, VOLS 1-8, 2007, : 1333 - 1336
  • [26] Model-based reinforcement learning: a computational model and an fMRI study
    Yoshida, W
    Ishii, S
    NEUROCOMPUTING, 2005, 63 : 253 - 269
  • [27] Combining Model-Based and Model-Free Reinforcement Learning Policies for More Efficient Sepsis Treatment
    Liu, Xiangyu
    Yu, Chao
    Huang, Qikai
    Wang, Luhao
    Wu, Jianfeng
    Guan, Xiangdong
    BIOINFORMATICS RESEARCH AND APPLICATIONS, ISBRA 2021, 2021, 13064 : 105 - 117
  • [28] Model-Based Reinforcement Learning for Cavity Filter Tuning
    Nimara, Doumitrou Daniil
    Malek-Mohammadi, Mohammadreza
    Wei, Jieqiang
    Huang, Vincent
    Ogren, Petter
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [29] Microblog Retrieval Based on Concept-Enhanced Pre-Training Model
    Wang, Yashen
    Wang, Zhaoyu
    Zhang, Huanhuan
    Liu, Zhirun
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2022, 17 (03)
  • [30] Model-based Bayesian Reinforcement Learning for Dialogue Management
    Lison, Pierre
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 475 - 479