An effective asynchronous framework for small scale reinforcement learning problems

被引:0
作者
Shifei Ding
Xingyu Zhao
Xinzheng Xu
Tongfeng Sun
Weikuan Jia
机构
[1] China University of Mining and Technology,School of Computer Science and Technology
[2] Mine DigitizationEngineering Research Center of Minstry of Education of the People′s Republic of China,School of Information Science and Engineering
[3] Shandong Normal University,undefined
来源
Applied Intelligence | 2019年 / 49卷
关键词
Reinforcement learning; Path planning; Asynchronous framework; Machine learning; Parallel framework;
D O I
暂无
中图分类号
学科分类号
摘要
Reinforcement learning is one of the research hotspots in the field of artificial intelligence in recent years. In the past few years, deep reinforcement learning has been widely used to solve various decision-making problems. However, due to the characteristics of neural networks, it is very easy to fall into local minima when facing small scale discrete space path planning problems. Traditional reinforcement learning uses continuous updating of a single agent when algorithm executes, which leads to a slow convergence speed. Although some scholars have done some improvement work to solve these problems, there are still many shortcomings to be overcome. In order to solve the above problems, we proposed a new asynchronous tabular reinforcement learning algorithms framework in this paper, and present four new variants of asynchronous reinforcement learning algorithms. We apply these algorithms on the standard reinforcement learning environments: frozen lake problem, cliff walking problem and windy gridworld problem, and the simulation results show that these methods can solve discrete space path planning problems efficiently and well balance the exploration and exploitation.
引用
收藏
页码:4303 / 4318
页数:15
相关论文
共 37 条
  • [1] Zhang Chongsheng(2019)Multi-Imbalance: An open-source software for multi-class imbalance learning Knowledge-Based Systems 174 137-143
  • [2] Bi Jingjun(2019)Computer Aided detection for fibrillations and flutters using deep convolutional neural network Inf Sci 486 231-239
  • [3] Xu Shixin(2017)An up-to-date comparison of state-of-the-art classification algorithms Expert Syst Appl 82 128-150
  • [4] Ramentol Enislay(2019)Multi-view manifold regularized learning-based method for prioritizing candidate disease miRNAs Knowledge-Based Systems 175 118-129
  • [5] Fan Gaojuan(2017)Mastering the game of Go without human knowledge Nature 550 354-359
  • [6] Qiao Baojun(2016)Mastering the game of Go with deep neural networks and tree search Nature 529 484-489
  • [7] Fujita Hamido(2015)Human-level control through deep reinforcement learning Nature 518 529-533
  • [8] Fujita H(1988)Learning to predict by the methods of temporal differences Mach Learn 3 9-44
  • [9] Cimr D(1994)Asynchronous stochastic approximation and Q-learning Mach Learn 16 185-202
  • [10] Zhang C(2018)Asynchronous Reinforcement Learning Algorithms for Solving Discrete Space Path Planning Problems Appl Intell 48 4889-4904