Generalization Enhancement of Visual Reinforcement Learning through Internal States

被引:0
作者
Yang, Hanlin [1 ]
Zhu, William [1 ]
Zhu, Xianchao [2 ]
机构
[1] Univ Elect Sci & Technol China, Inst Fundamental & Frontier Sci, Chengdu 611731, Peoples R China
[2] Henan Univ Technol, Sch Artificial Intelligence & Big Data, Zhengzhou 450001, Peoples R China
关键词
visual reinforcement learning; transfer learning; generalization;
D O I
10.3390/s24144513
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Visual reinforcement learning is important in various practical applications, such as video games, robotic manipulation, and autonomous navigation. However, a major challenge in visual reinforcement learning is the generalization to unseen environments, that is, how agents manage environments with previously unseen backgrounds. This issue is triggered mainly by the high unpredictability inherent in high-dimensional observation space. To deal with this problem, techniques including domain randomization and data augmentation have been explored; nevertheless, these methods still cannot attain a satisfactory result. This paper proposes a new method named Internal States Simulation Auxiliary (ISSA), which uses internal states to improve generalization in visual reinforcement learning tasks. Our method contains two agents, a teacher agent and a student agent: the teacher agent has the ability to directly access the environment's internal states and is used to facilitate the student agent's training; the student agent receives initial guidance from the teacher agent and subsequently continues to learn independently. From another perspective, our method can be divided into two phases, the transfer learning phase and traditional visual reinforcement learning phase. In the first phase, the teacher agent interacts with environments and imparts knowledge to the vision-based student agent. With the guidance of the teacher agent, the student agent is able to discover more effective visual representations that address the high unpredictability of high-dimensional observation space. In the next phase, the student agent autonomously learns from the visual information in the environment, and ultimately, it becomes a vision-based reinforcement learning agent with enhanced generalization. The effectiveness of our method is evaluated using the DMControl Generalization Benchmark and the DrawerWorld with texture distortions. Preliminary results indicate that our method significantly improves generalization ability and performance in complex continuous control tasks.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Enhanced image steganalysis through reinforcement learning and generative adversarial networks
    Al-Obaidi, Sumia Abdulhussien Razooqi
    Lighvan, Mina Zolfy
    Asadpour, Mohammad
    INTELLIGENT DECISION TECHNOLOGIES-NETHERLANDS, 2024, 18 (02): : 1077 - 1100
  • [42] Improving generalization to new environments and removing catastrophic forgetting in Reinforcement Learning by using an eco-system of agents
    Moulin, Olivier
    Vincent-Laver, Francois
    Elbers, Paul
    Hoogendoorn, Mark
    2022 IEEE/WIC/ACM INTERNATIONAL JOINT CONFERENCE ON WEB INTELLIGENCE AND INTELLIGENT AGENT TECHNOLOGY, WI-IAT, 2022, : 166 - 173
  • [43] Robot gaining accurate pouring skills through self-supervised learning and generalization
    Huang, Yongqiang
    Wilches, Juan
    Sun, Yu
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2021, 136
  • [44] Enhancing Reinforcement Learning-Based Energy Management Through Transfer Learning With Load and PV Forecasting
    Xu, Chang
    Inuiguchi, Masahiro
    Hayashi, Naoki
    Raymond, Wong Jee Keen
    Mokhlis, Hazlie
    Illias, Hazlee Azil
    IEEE ACCESS, 2025, 13 : 43956 - 43972
  • [45] Bridging the Reality Gap Between Virtual and Physical Environments Through Reinforcement Learning
    Ranaweera, Mahesh
    Mahmoud, Qusay H.
    IEEE ACCESS, 2023, 11 : 19914 - 19927
  • [46] Data Augmentation in Latent Space with Variational Autoencoder and Pretrained Image Model for Visual Reinforcement Learning
    Dang, Xuzhe
    Edelkamp, Stefan
    KI 2024: ADVANCES IN ARTIFICIAL INTELLIGENCE, KI 2024, 2024, 14992 : 45 - 59
  • [47] Don't overlook any detail: Data-efficient reinforcement learning with visual attention
    Ma, Jialin
    Li, Ce
    Feng, Zhiqiang
    Xiao, Limei
    He, Chengdan
    Zhang, Yan
    KNOWLEDGE-BASED SYSTEMS, 2025, 310
  • [48] Synthesizing CT images from MR images with deep learning: model generalization for different datasets through transfer learning
    Li, Wen
    Kazemifar, Samaneh
    Bai, Ti
    Nguyen, Dan
    Weng, Yaochung
    Li, Yafen
    Xia, Jun
    Xiong, Jing
    Xie, Yaoqin
    Owrangi, Amir
    Jiang, Steve
    BIOMEDICAL PHYSICS & ENGINEERING EXPRESS, 2021, 7 (02)
  • [49] TraKDis: A Transformer-Based Knowledge Distillation Approach for Visual Reinforcement Learning With Application to Cloth Manipulation
    Chen, Wei
    Rojas, Nicolas
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (03) : 2455 - 2462
  • [50] Temporal goal-aware transformer assisted visual reinforcement learning for virtual table tennis agent
    Wang, Jinyang
    Wang, Jihong
    Li, Haoxuan
    Huang, Xiaojun
    Xia, Jun
    Li, Zhen
    Wu, Weibing
    Sheng, Bin
    VISUAL COMPUTER, 2025,