Network parameter setting for reinforcement learning approaches using neural networks

被引:0
作者
Yamada, Kazuaki [1 ]
Ohkura, Kazuhiro [1 ]
机构
[1] Department of Mechanical Engineering, Undergraduate School of Science and Techonology, Toyo University, Kawagoe-shi, Saitama, 350-8585, 2100, Kujirai
来源
Nihon Kikai Gakkai Ronbunshu, C Hen/Transactions of the Japan Society of Mechanical Engineers, Part C | 2012年 / 78卷 / 792期
关键词
Autonomous mobile robot; Neural networks; Reinforcement learning;
D O I
10.1299/kikaic.78.2950
中图分类号
学科分类号
摘要
Reinforcement learning approaches attract attention as the technique to construct the mapping function between sensors-motors of an autonomous robot through trial-and-error. Traditional reinforcement learning approaches make use of look-up table to express the mapping function between the grid state space and the grid action space. However the grid size of the state space affects the learning performances significantly. To overcome this problem, many researchers have proposed algorithms using neural networks to express the mapping function between the continuous state space and actions. However, in this case, a designer needs to appropriately set the number of middle neurons and the initial value of weight parameters of neural networks to improve the approximate accuracy of neural networks. This paper proposes a new method to automatically set the number of middle neurons and the initial value of the weight parameters of neural networks, on the basis of the dimensional-number of the sensor space, in Q-learning using neural networks. The proposed method is demonstrated through a navigation problem of an autonomous mobile robot, and is evaluated by comparing Q-learning using RBF networks and Q-learning using neural networks whose parameters are set by a designer. © 2012 The Japan Society of Mechanical Engineers.
引用
收藏
页码:2950 / 2961
页数:11
相关论文
共 50 条
  • [31] Data-Driven Motion Planning: A Survey on Deep Neural Networks, Reinforcement Learning, and Large Language Model Approaches
    de Carvalho, Gabriel Peixoto
    Sawanobori, Tetsuya
    Horii, Takato
    IEEE ACCESS, 2025, 13 : 52195 - 52245
  • [32] Stable reinforcement learning with recurrent neural networks
    James Nate KNIGHT
    Charles ANDERSON
    JournalofControlTheoryandApplications, 2011, 9 (03) : 410 - 420
  • [33] Stable reinforcement learning with recurrent neural networks
    Knight J.N.
    Anderson C.
    Journal of Control Theory and Applications, 2011, 9 (3): : 410 - 420
  • [34] Reinforcement Learning of Graph Neural Networks for Service Function Chaining in Computer Network Management
    Heo, DongNyeong
    Lee, Doyoung
    Kim, Hee-Gon
    Park, Suhyun
    Choi, Heeyoul
    2022 23RD ASIA-PACIFIC NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM (APNOMS 2022), 2022, : 139 - 144
  • [35] Neural network compression for reinforcement learning tasks
    Dmitry A. Ivanov
    Denis A. Larionov
    Oleg V. Maslennikov
    Vladimir V. Voevodin
    Scientific Reports, 15 (1)
  • [36] Parameter Setting for Deep Neural Networks Using Swarm Intelligence on Phishing Websites Classification
    Vrbancic, Grega
    Fister, Iztok, Jr.
    Podgorelec, Vili
    INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2019, 28 (06)
  • [37] Reinforcement learning with internal expectation in the random neural networks for cascaded decisions
    Halici, U
    BIOSYSTEMS, 2001, 63 (1-3) : 21 - 34
  • [38] Identifiability in training neural networks for reconfigurable control based on reinforcement learning
    de Weerdt, E.
    Chu, Q. P.
    Mulder, J. A.
    PROCEEDINGS OF THE NINTH IASTED INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND CONTROL, 2006, : 7 - +
  • [39] Optimizing coverage in wireless sensor networks using deep reinforcement learning with graph neural networks
    G. Pushpa
    R. Anand Babu
    S. Subashree
    S. Senthilkumar
    Scientific Reports, 15 (1)
  • [40] Obstacle avoidance of redundant manipulators using neural networks based reinforcement learning
    Duguleana, Mihai
    Barbuceanu, Florin Grigore
    Teirelbar, Ahmed
    Mogan, Gheorghe
    ROBOTICS AND COMPUTER-INTEGRATED MANUFACTURING, 2012, 28 (02) : 132 - 146