Network parameter setting for reinforcement learning approaches using neural networks

被引:0
作者
Yamada, Kazuaki [1 ]
Ohkura, Kazuhiro [1 ]
机构
[1] Department of Mechanical Engineering, Undergraduate School of Science and Techonology, Toyo University, Kawagoe-shi, Saitama, 350-8585, 2100, Kujirai
来源
Nihon Kikai Gakkai Ronbunshu, C Hen/Transactions of the Japan Society of Mechanical Engineers, Part C | 2012年 / 78卷 / 792期
关键词
Autonomous mobile robot; Neural networks; Reinforcement learning;
D O I
10.1299/kikaic.78.2950
中图分类号
学科分类号
摘要
Reinforcement learning approaches attract attention as the technique to construct the mapping function between sensors-motors of an autonomous robot through trial-and-error. Traditional reinforcement learning approaches make use of look-up table to express the mapping function between the grid state space and the grid action space. However the grid size of the state space affects the learning performances significantly. To overcome this problem, many researchers have proposed algorithms using neural networks to express the mapping function between the continuous state space and actions. However, in this case, a designer needs to appropriately set the number of middle neurons and the initial value of weight parameters of neural networks to improve the approximate accuracy of neural networks. This paper proposes a new method to automatically set the number of middle neurons and the initial value of the weight parameters of neural networks, on the basis of the dimensional-number of the sensor space, in Q-learning using neural networks. The proposed method is demonstrated through a navigation problem of an autonomous mobile robot, and is evaluated by comparing Q-learning using RBF networks and Q-learning using neural networks whose parameters are set by a designer. © 2012 The Japan Society of Mechanical Engineers.
引用
收藏
页码:2950 / 2961
页数:11
相关论文
共 50 条
  • [41] A working memory model based on recurrent neural networks using reinforcement learning
    Wang, Mengyuan
    Wang, Yihong
    Xu, Xuying
    Pan, Xiaochuan
    [J]. COGNITIVE NEURODYNAMICS, 2024, 18 (05) : 3031 - 3058
  • [42] Autonomous 3-D UAV Localization Using Cellular Networks: Deep Supervised Learning Versus Reinforcement Learning Approaches
    Afifi, Ghada
    Gadallah, Yasser
    [J]. IEEE ACCESS, 2021, 9 : 155234 - 155248
  • [43] Parameter incremental learning algorithm for neural networks
    Wan, Sheng
    Banta, Larry E.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2006, 17 (06): : 1424 - 1438
  • [44] Combination of Neural Networks and Reinforcement Learning for Wind Turbine Pitch Control
    Sierra-Garcia, Jesus Enrique
    Santos, Matilde
    [J]. HYBRID ARTIFICIAL INTELLIGENT SYSTEMS, HAIS 2022, 2022, 13469 : 385 - 392
  • [45] Multi-Task Reinforcement Meta-Learning in Neural Networks
    Shakah, Ghazi
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (07) : 263 - 269
  • [46] Network Intrusion Detection Based on Extended RBF Neural Network With Offline Reinforcement Learning
    Lopez-Martin, Manuel
    Sanchez-Esguevillas, Antonio
    Ignacio Arribas, Juan
    Carro, Belen
    [J]. IEEE ACCESS, 2021, 9 : 153153 - 153170
  • [47] Mobile Robot Heading Adjustment Using Radial Basis Function Neural Networks Controller and Reinforcement Learning
    Bayar, Goekhan
    Konukseven, E. Ilhan
    Koku, A. Bugra
    [J]. PROCEEDINGS OF THE 4TH WSEAS/IASME INTERNATIONAL CONFERENCE ON DYNAMICAL SYSTEMS AND CONTROLS, 2008, : 169 - 174
  • [48] Reinforcement learning algorithm with network extension for pulse neural network
    Takita, M
    Osana, Y
    Hagiwara, M
    [J]. SMC 2000 CONFERENCE PROCEEDINGS: 2000 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN & CYBERNETICS, VOL 1-5, 2000, : 2586 - 2591
  • [49] Reinforcement learning using the stochastic fuzzy min-max neural network
    Likas, A
    [J]. NEURAL PROCESSING LETTERS, 2001, 13 (03) : 213 - 220
  • [50] Distantly supervised biomedical relation extraction using piecewise attentive convolutional neural network and reinforcement learning
    Zhu, Tiantian
    Qin, Yang
    Xiang, Yang
    Hu, Baotian
    Chen, Qingcai
    Peng, Weihua
    [J]. JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2021, 28 (12) : 2571 - 2581