Multiresolution state-space discretization for Q-Learning with pseudorandomized discretization

被引:3
作者
Lampton A. [1 ]
Valasek J. [2 ]
Kumar M. [3 ]
机构
[1] Systems Technology, Inc., Hawthorne, CA 90250
[2] Department of Aerospace Engineering, Texas A and M University, College Station, TX 77843-3141
[3] Department of Mechanical and Aerospace Engineering, University of Florida, Gainesville, FL 32611-6250
来源
Journal of Control Theory and Applications | 2011年 / 9卷 / 3期
基金
美国国家科学基金会;
关键词
Morphing; Random grid; Reinforcement learning;
D O I
10.1007/s11768-011-1012-4
中图分类号
学科分类号
摘要
A multiresolution state-space discretization method with pseudorandom gridding is developed for the episodic unsupervised learning method of Q-learning. It is used as the learning agent for closed-loop control of morphing or highly reconfigurable systems. This paper develops a method whereby a state-space is adaptively discretized by progressively finer pseudorandom grids around the regions of interest within the state or learning space in an effort to break the Curse of Dimensionality. Utility of the method is demonstrated with application to the problem of a morphing airfoil, which is simulated by a computationally intensive computational fluid dynamics model. By setting the multiresolution method to define the region of interest by the goal the agent seeks, it is shown that this method with the pseudorandom grid can learn a specific goal within ±0.001 while reducing the total number of state-action pairs needed to achieve this level of specificity to less than 3000. © 2011 South China University of Technology, Academy of Mathematics and Systems Science, Chinese Academy of Sciences and Springer-Verlag Berlin Heidelberg.
引用
收藏
页码:431 / 439
页数:8
相关论文
共 34 条
  • [21] Q-LEARNING BY THE nth STEP STATE AND MULTI-AGENT NEGOTIATION IN UNKNOWN ENVIRONMENT
    Job, Josip
    Jovic, Franjo
    Livada, Caslav
    TEHNICKI VJESNIK-TECHNICAL GAZETTE, 2012, 19 (03): : 529 - 534
  • [22] Hierarchical state-abstracted and socially augmented Q-Learning for reducing complexity in agent-based learning
    Laura RAY
    Journal of Control Theory and Applications, 2011, 9 (03) : 440 - 450
  • [23] Hierarchical state-abstracted and socially augmented Q-Learning for reducing complexity in agent-based learning
    Sun X.
    Mao T.
    Ray L.
    Shi D.
    Kralik J.
    Journal of Control Theory and Applications, 2011, 9 (03): : 440 - 450
  • [24] Improved Q-Learning Algorithm Based on Approximate State Matching in Agricultural Plant Protection Environment
    Sun, Fengjie
    Wang, Xianchang
    Zhang, Rui
    ENTROPY, 2021, 23 (06)
  • [25] Multi-objective fuzzy Q-learning to solve continuous state-action problems
    Asgharnia, Amirhossein
    Schwartz, Howard
    Atia, Mohamed
    NEUROCOMPUTING, 2023, 516 : 115 - 132
  • [26] State-Space Closure: Revisiting Endless Online Level Generation via Reinforcement Learning
    Wang, Ziqi
    Shu, Tianye
    Liu, Jialin
    IEEE TRANSACTIONS ON GAMES, 2024, 16 (02) : 489 - 492
  • [27] STATE-SPACE CHARACTERIZATION OF HUMAN BALANCE THROUGH A REINFORCEMENT LEARNING BASED MUSCLE CONTROLLER
    Akbas, Kubra
    Zhou, Xianlian
    PROCEEDINGS OF ASME 2023 INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, IDETC-CIE2023, VOL 2, 2023,
  • [28] Online Gaussian Process State-space Model: Learning and Planning for Partially Observable Dynamical Systems
    Soon-Seo Park
    Young-Jin Park
    Youngjae Min
    Han-Lim Choi
    International Journal of Control, Automation and Systems, 2022, 20 : 601 - 617
  • [29] Potential-based reward shaping using state-space segmentation for efficiency in reinforcement learning
    Bal, Melis Ilayda
    Aydin, Hueseyin
    Iyiguen, Cem
    Polat, Faruk
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 157 : 469 - 484
  • [30] Online Gaussian Process State-space Model: Learning and Planning for Partially Observable Dynamical Systems
    Park, Soon-Seo
    Park, Young-Jin
    Min, Youngjae
    Choi, Han-Lim
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2022, 20 (02) : 601 - 617