Network parameter setting for reinforcement learning approaches using neural networks

被引:0
|
作者
Yamada, Kazuaki [1 ]
Ohkura, Kazuhiro [1 ]
机构
[1] Department of Mechanical Engineering, Undergraduate School of Science and Techonology, Toyo University, Kawagoe-shi, Saitama, 350-8585, 2100, Kujirai
来源
Nihon Kikai Gakkai Ronbunshu, C Hen/Transactions of the Japan Society of Mechanical Engineers, Part C | 2012年 / 78卷 / 792期
关键词
Autonomous mobile robot; Neural networks; Reinforcement learning;
D O I
10.1299/kikaic.78.2950
中图分类号
学科分类号
摘要
Reinforcement learning approaches attract attention as the technique to construct the mapping function between sensors-motors of an autonomous robot through trial-and-error. Traditional reinforcement learning approaches make use of look-up table to express the mapping function between the grid state space and the grid action space. However the grid size of the state space affects the learning performances significantly. To overcome this problem, many researchers have proposed algorithms using neural networks to express the mapping function between the continuous state space and actions. However, in this case, a designer needs to appropriately set the number of middle neurons and the initial value of weight parameters of neural networks to improve the approximate accuracy of neural networks. This paper proposes a new method to automatically set the number of middle neurons and the initial value of the weight parameters of neural networks, on the basis of the dimensional-number of the sensor space, in Q-learning using neural networks. The proposed method is demonstrated through a navigation problem of an autonomous mobile robot, and is evaluated by comparing Q-learning using RBF networks and Q-learning using neural networks whose parameters are set by a designer. © 2012 The Japan Society of Mechanical Engineers.
引用
收藏
页码:2950 / 2961
页数:11
相关论文
共 50 条
  • [21] Learning strategy with neural-networks and reinforcement learning for actual manipulator robot
    Nakamura, Shingo
    Hashimoto, Shuji
    PROCEEDINGS OF THE SEVENTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 17TH '12), 2012, : 947 - 950
  • [22] Reinforcement Learning Using the Stochastic Fuzzy Min–Max Neural Network
    Aristidis Likas
    Neural Processing Letters, 2001, 13 : 213 - 220
  • [23] Category learning in a recurrent neural network with reinforcement learning
    Zhang, Ying
    Pan, Xiaochuan
    Wang, Yihong
    FRONTIERS IN PSYCHIATRY, 2022, 13
  • [24] Emergence of Higher Exploration in Reinforcement Learning Using a Chaotic Neural Network
    Goto, Yuki
    Shibata, Katsunari
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 40 - 48
  • [25] A Hybrid Neural Network Model Based Reinforcement Learning Agent
    Gao, Pengyi
    Chen, Chuanbo
    Zhang, Kui
    Hu, Yingsong
    Li, Dan
    ADVANCES IN NEURAL NETWORKS - ISNN 2010, PT 1, PROCEEDINGS, 2010, 6063 : 436 - +
  • [26] Reinforcement Learning in Card Game Environments Using Monte Carlo Methods and Artificial Neural Networks
    Baykal, Omer
    Alpaslan, Ferda Nur
    2019 4TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND ENGINEERING (UBMK), 2019, : 618 - 623
  • [27] USING MODULAR NEURAL NETWORKS AND MACHINE LEARNING WITH REINFORCEMENT LEARNING TO SOLVE CLASSIFICATION PROBLEMS
    Leoshchenko, S. D.
    Oliinyk, A. O.
    Subbotin, S. A.
    Kolpakova, T. O.
    RADIO ELECTRONICS COMPUTER SCIENCE CONTROL, 2024, (02) : 71 - 81
  • [28] Influence of the Chaotic Property on Reinforcement Learning Using a Chaotic Neural Network
    Goto, Yuki
    Shibata, Katsunari
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT I, 2017, 10634 : 759 - 767
  • [29] Mobile Robot Navigation Using Reinforcement Learning Based on Neural Network with Short Term Memory
    Gavrilov, Andrey V.
    Lenskiy, Artem
    ADVANCED INTELLIGENT COMPUTING, 2011, 6838 : 210 - +
  • [30] Data-Driven Motion Planning: A Survey on Deep Neural Networks, Reinforcement Learning, and Large Language Model Approaches
    de Carvalho, Gabriel Peixoto
    Sawanobori, Tetsuya
    Horii, Takato
    IEEE ACCESS, 2025, 13 : 52195 - 52245