Improving Reinforcement Learning Exploration by Autoencoders

被引:0
|
作者
Paczolay, Gabor [1 ]
Harmati, Istvan [1 ]
机构
[1] Department of Control Engineering, Budapest University of Technology and Economics, Magyar Tudósok körútja 2., Budapest
来源
Periodica Polytechnica Electrical Engineering and Computer Science | 2024年 / 68卷 / 04期
关键词
AutE-DQN; autoencoders; DQN; exploration; reinforcement learning;
D O I
10.3311/PPee.36789
中图分类号
学科分类号
摘要
Reinforcement learning is a field with massive potential related to solving engineering problems without field knowledge. However, the problem of exploration and exploitation emerges when one tries to balance a system between the learning phase and proper execution. In this paper, a new method is proposed that utilizes autoencoders to manage the exploration rate in an epsilon-greedy exploration algorithm. The error between the real state and the reconstructed state by the autoencoder becomes the base of the exploration-exploitation rate. The proposed method is then examined in two experiments: one benchmark is the cartpole experiment while the other is a gridworld example created for this paper to examine long-term exploration. Both experiments show results such that the proposed method performs better in these scenarios. © 2024 Budapest University of Technology and Economics. All rights reserved.
引用
收藏
页码:335 / 343
页数:8
相关论文
共 50 条
  • [1] Improving Deep Reinforcement Learning With Transitional Variational Autoencoders: A Healthcare Application
    Baucum, Matthew
    Khojandi, Anahita
    Vasudevan, Rama
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (06) : 2273 - 2280
  • [2] Improving exploration in deep reinforcement learning for stock trading
    Zemzem, Wiem
    Tagina, Moncef
    INTERNATIONAL JOURNAL OF COMPUTER APPLICATIONS IN TECHNOLOGY, 2023, 72 (04) : 288 - 295
  • [3] On the use of Deep Autoencoders for Efficient Embedded Reinforcement Learning
    Prakash, Bharat
    Horton, Mark
    Waytowich, Nicholas R.
    Hairston, William David
    Oates, Tim
    Mohsenin, Tinoosh
    GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 507 - 512
  • [4] Learning-Driven Exploration for Reinforcement Learning
    Usama, Muhammad
    Chang, Dong Eui
    2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 1146 - 1151
  • [5] Adaptive Exploration Strategies for Reinforcement Learning
    Hwang, Kao-Shing
    Li, Chih-Wen
    Jiang, Wei-Cheng
    2017 INTERNATIONAL CONFERENCE ON SYSTEM SCIENCE AND ENGINEERING (ICSSE), 2017, : 16 - 19
  • [6] Reinforcement Learning with Derivative-Free Exploration
    Chen, Xiong-Hui
    Yu, Yang
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 1880 - 1882
  • [7] Exploration With Task Information for Meta Reinforcement Learning
    Jiang, Peng
    Song, Shiji
    Huang, Gao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4033 - 4046
  • [8] Intrinsically Motivated Lifelong Exploration in Reinforcement Learning
    Bougie, Nicolas
    Ichise, Ryutaro
    ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 1357 : 109 - 120
  • [9] Learning to soar: Resource-constrained exploration in reinforcement learning
    Chung, Jen Jen
    Lawrance, Nicholas R. J.
    Sukkarieh, Salah
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2015, 34 (02) : 158 - 172
  • [10] Exploration and Incentives in Reinforcement Learning
    Simchowitz, Max
    Slivkins, Aleksandrs
    OPERATIONS RESEARCH, 2024, 72 (03) : 983 - 998