GRIMGEP: Learning Progress for Robust Goal Sampling in Visual Deep Reinforcement Learning

被引:2
|
作者
Kovac, Grgur [1 ]
Laversanne-Finot, Adrien [1 ]
Oudeyer, Pierre-Yves [1 ]
机构
[1] INRIA Bordeaux, Flowers Lab, F-33400 Talence, France
关键词
Goal exploration; learning progress; reinforcement learning (RL); INTRINSIC MOTIVATION; EXPLORATION; SYSTEMS;
D O I
10.1109/TCDS.2022.3216911
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Autotelic reinforcement learning (RL) agents sample their own goals, and try to reach them. They often prioritize goal sampling according to some intrinsic reward, ex. novelty or absolute learning progress (ALPs). Novelty-based approaches work robustly in unsupervised image-based environments when there are no distractors. However, they construct simple curricula that do not take the agent's performance into account: in complex environments, they often get attracted by impossible tasks. ALP-based approaches, which are often combined with a clustering mechanism, construct complex curricula tuned to the agent's current capabilities. Such curricula sample goals on which the agent is currently learning the most, and do not get attracted by impossible tasks. However, ALP approaches have not so far been applied to DRL agents perceiving complex environments directly in the image space. Goal regions guided intrinsically motivated goal exploration process (GRIMGEP), without using any expert knowledge, combines the ALP clustering approaches with novelty-based approaches and extends them to those complex scenarios. We experiment on a rich 3-D image-based environment with distractors using novelty-based exploration approaches: Skewfit and CountBased. We show that wrapping them with GRIMGEP-using them only in the cluster sampled by ALP-creates a better curriculum. The wrapped approaches are attracted less by the distractors, and achieve drastically better performances.
引用
收藏
页码:1396 / 1407
页数:12
相关论文
共 50 条
  • [1] A Deep Reinforcement Learning Approach to Configuration Sampling Problem
    Abolfazli, Amir
    Spiegetberg, Jakob
    Palmer, Gregory
    Anand, Avishek
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 1 - 10
  • [2] Deep Reinforcement Learning For Visual Navigation of Wheeled Mobile Robots
    Nwaonumah, Ezebuugo
    Samanta, Biswanath
    IEEE SOUTHEASTCON 2020, 2020,
  • [3] Exploration in deep reinforcement learning: A survey
    Ladosz, Pawel
    Weng, Lilian
    Kim, Minwoo
    Oh, Hyondong
    INFORMATION FUSION, 2022, 85 : 1 - 22
  • [4] Robust visual tracking based on scale invariance and deep learning
    Ren, Nan
    Du, Junping
    Zhu, Suguo
    Li, Linghui
    Fan, Dan
    Lee, JangMyung
    FRONTIERS OF COMPUTER SCIENCE, 2017, 11 (02) : 230 - 242
  • [5] Action-Driven Visual Object Tracking With Deep Reinforcement Learning
    Yun, Sangdoo
    Choi, Jongwon
    Yoo, Youngjoon
    Yun, Kimin
    Choi, Jin Young
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) : 2239 - 2252
  • [6] Automatic Goal Generation for Reinforcement Learning Agents
    Florensa, Carlos
    Held, David
    Geng, Xinyang
    Abbeel, Pieter
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [7] Testing Theories of Goal Progress in Online Learning
    Lu, Joy
    Bradlow, Eric T.
    Hutchinson, J. Wesley
    JOURNAL OF MARKETING RESEARCH, 2022, 59 (01) : 35 - 60
  • [8] Adaptive and Robust Network Routing Based on Deep Reinforcement Learning with Lyapunov Optimization
    Zhuang, Zirui
    Wang, Jingyu
    Qi, Qi
    Liao, Jianxin
    Han, Zhu
    2020 IEEE/ACM 28TH INTERNATIONAL SYMPOSIUM ON QUALITY OF SERVICE (IWQOS), 2020,
  • [9] Multitask Learning for Object Localization With Deep Reinforcement Learning
    Wang, Yan
    Zhang, Lei
    Wang, Lituan
    Wang, Zizhou
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2019, 11 (04) : 573 - 580
  • [10] Visual Active Tracking Algorithm for UAV Cluster Based on Deep Reinforcement Learning
    Hu, Runqiao
    Wang, Shaofan
    Li, Ke
    PROCEEDINGS OF 2022 INTERNATIONAL CONFERENCE ON AUTONOMOUS UNMANNED SYSTEMS, ICAUS 2022, 2023, 1010 : 1047 - 1061