Pre-training with Augmentations for Efficient Transfer in Model-Based Reinforcement Learning

被引:0
|
作者
Esteves, Bernardo [1 ,2 ]
Vasco, Miguel [1 ,2 ]
Melo, Francisco S. [1 ,2 ]
机构
[1] INESC ID, Lisbon, Portugal
[2] Univ Lisbon, Inst Super Tecn, Lisbon, Portugal
来源
PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2023, PT I | 2023年 / 14115卷
关键词
Reinforcement learning; Transfer learning; Representation learning;
D O I
10.1007/978-3-031-49008-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This work explores pre-training as a strategy to allow reinforcement learning (RL) algorithms to efficiently adapt to new (albeit similar) tasks. We argue for introducing variability during the pre-training phase, in the form of augmentations to the observations of the agent, to improve the sample efficiency of the fine-tuning stage. We categorize such variability in the form of perceptual, dynamic and semantic augmentations, which can be easily employed in standard pre-training methods. We perform extensive evaluations of our proposed augmentation scheme in model-based algorithms, across multiple scenarios of increasing complexity. The results consistently show that our augmentation scheme significantly improves the efficiency of the fine-tuning to novel tasks, outperforming other state-of-the-art pre-training approaches.
引用
收藏
页码:133 / 145
页数:13
相关论文
共 50 条
  • [41] Multi-task Pre-training with Soft Biometrics for Transfer-learning Palmprint Recognition
    Xu, Huanhuan
    Leng, Lu
    Yang, Ziyuan
    Teoh, Andrew Beng Jin
    Jin, Zhe
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 2341 - 2358
  • [42] Case-Based Task Generalization in Model-Based Reinforcement Learning
    Zholus, Artem
    Panov, Aleksandr, I
    ARTIFICIAL GENERAL INTELLIGENCE, AGI 2021, 2022, 13154 : 344 - 354
  • [43] Sequential Monte Carlo Samplers for Model-Based Reinforcement Learning
    Sonmez, Orhan
    Cemgil, A. Taylan
    2013 21ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2013,
  • [44] Contrastive State Augmentations for Reinforcement Learning-Based Recommender Systems
    Ren, Zhaochun
    Huang, Na
    Wang, Yidan
    Ren, Pengjie
    Ma, Jun
    Lei, Jiahuan
    Shi, Xinlei
    Luo, Hengliang
    Jose, Joemon
    Xin, Xin
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 922 - 931
  • [45] Model-Based Reinforcement Learning via Stochastic Hybrid Models
    Abdulsamad, Hany
    Peters, Jan
    IEEE OPEN JOURNAL OF CONTROL SYSTEMS, 2023, 2 : 155 - 170
  • [46] Model-Based Offline Reinforcement Learning for Autonomous Delivery of Guidewire
    Li, Hao
    Zhou, Xiao-Hu
    Xie, Xiao-Liang
    Liu, Shi-Qi
    Feng, Zhen-Qiu
    Gui, Mei-Jiang
    Xiang, Tian-Yu
    Huang, De-Xing
    Hou, Zeng-Guang
    IEEE TRANSACTIONS ON MEDICAL ROBOTICS AND BIONICS, 2024, 6 (03): : 1054 - 1062
  • [47] Federated Ensemble Model-Based Reinforcement Learning in Edge Computing
    Wang, Jin
    Hu, Jia
    Mills, Jed
    Min, Geyong
    Xia, Ming
    Georgalas, Nektarios
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (06) : 1848 - 1859
  • [48] Model-based hierarchical reinforcement learning and human action control
    Botvinick, Matthew
    Weinstein, Ari
    PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2014, 369 (1655)
  • [49] Model-based Reinforcement Learning for Ship Path Following with Disturbances
    Dong, Zhengyang
    Chen, Linying
    Huang, Yamin
    Chen, Pengfei
    Mou, Junmin
    IFAC PAPERSONLINE, 2024, 58 (20): : 247 - 252
  • [50] Intrinsic Motivation in Model-Based Reinforcement Learning: A Brief Review
    A. K. Latyshev
    A. I. Panov
    Scientific and Technical Information Processing, 2024, 51 (5) : 460 - 470