Pre-training with Augmentations for Efficient Transfer in Model-Based Reinforcement Learning

被引:0
|
作者
Esteves, Bernardo [1 ,2 ]
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
机构
[1] INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
来源
PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2023, PT I | 2023年 / 14115卷
关键词
Reinforcement learning; Transfer learning; Representation learning;
D O I
10.1007/978-3-031-49008-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work explores pre-training as a strategy to allow reinforcement learning (RL) algorithms to efficiently adapt to new (albeit similar) tasks. We argue for introducing variability during the pre-training phase, in the form of augmentations to the observations of the agent, to improve the sample efficiency of the fine-tuning stage. We categorize such variability in the form of perceptual, dynamic and semantic augmentations, which can be easily employed in standard pre-training methods. We perform extensive evaluations of our proposed augmentation scheme in model-based algorithms, across multiple scenarios of increasing complexity. The results consistently show that our augmentation scheme significantly improves the efficiency of the fine-tuning to novel tasks, outperforming other state-of-the-art pre-training approaches.
引用
收藏
页码:133 / 145
页数:13
相关论文
共 50 条
  • [31] Model-based reinforcement learning for alternating Markov games
    Mellor, D
    AI 2003: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2003, 2903 : 520 - 531
  • [32] Learnable Weighting Mechanism in Model-based Reinforcement Learning
    Huang W.-Z.
    Yin Q.-Y.
    Zhang J.-G.
    Huang K.-Q.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (06): : 2765 - 2775
  • [33] Exploration in Relational Domains for Model-based Reinforcement Learning
    Lang, Tobias
    Toussaint, Marc
    Kersting, Kristian
    JOURNAL OF MACHINE LEARNING RESEARCH, 2012, 13 : 3725 - 3768
  • [34] A Model-based Factored Bayesian Reinforcement Learning Approach
    Wu, Bo
    Feng, Yanpeng
    Zheng, Hongyan
    APPLIED SCIENCE, MATERIALS SCIENCE AND INFORMATION TECHNOLOGIES IN INDUSTRY, 2014, 513-517 : 1092 - 1095
  • [35] Free Will Belief as a Consequence of Model-Based Reinforcement Learning
    Rehn, Erik M.
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2022, 2023, 13539 : 353 - 363
  • [36] Offline Model-Based Reinforcement Learning for Tokamak Control
    Char, Ian
    Abbate, Joseph
    Bardoczi, Laszlo
    Boyer, Mark D.
    Chung, Youngseog
    Conlin, Rory
    Erickson, Keith
    Mehta, Viraj
    Richner, Nathan
    Kolemen, Egemen
    Schneider, Jeff
    LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 211, 2023, 211
  • [37] R x R: Rapid eXploration for Reinforcement learning via sampling-based reset distributions and imitation pre-training
    Khandate, Gagan
    Saidi, Tristan L.
    Shang, Siqi
    Chang, Eric T.
    Liu, Yang
    Dennis, Seth
    Adams, Johnson
    Ciocarlie, Matei
    AUTONOMOUS ROBOTS, 2024, 48 (07)
  • [38] Pre-training Strategies and Datasets for Facial Representation Learning
    Bulat, Adrian
    Cheng, Shiyang
    Yang, Jing
    Garbett, Andrew
    Sanchez, Enrique
    Tzimiropoulos, Georgios
    COMPUTER VISION, ECCV 2022, PT XIII, 2022, 13673 : 107 - 125
  • [39] Data-Efficient Task Generalization via Probabilistic Model-Based Meta Reinforcement Learning
    Bhardwaj, Arjun
    Rothfuss, Jonas
    Sukhija, Bhavya
    As, Yarden
    Hutter, Marco
    Coros, Stelian
    Krause, Andreas
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (04) : 3918 - 3925
  • [40] Multi-task Pre-training with Soft Biometrics for Transfer-learning Palmprint Recognition
    Huanhuan Xu
    Lu Leng
    Ziyuan Yang
    Andrew Beng Jin Teoh
    Zhe Jin
    Neural Processing Letters, 2023, 55 : 2341 - 2358