Multiresolution state-space discretization for Q-Learning with pseudorandomized discretization

被引:3
|
作者
Lampton A. [1 ]
Valasek J. [2 ]
Kumar M. [3 ]
机构
[1] Systems Technology, Inc., Hawthorne, CA 90250
[2] Department of Aerospace Engineering, Texas A and M University, College Station, TX 77843-3141
[3] Department of Mechanical and Aerospace Engineering, University of Florida, Gainesville, FL 32611-6250
来源
Journal of Control Theory and Applications | 2011年 / 9卷 / 3期
基金
美国国家科学基金会;
关键词
Morphing; Random grid; Reinforcement learning;
D O I
10.1007/s11768-011-1012-4
中图分类号
学科分类号
摘要
A multiresolution state-space discretization method with pseudorandom gridding is developed for the episodic unsupervised learning method of Q-learning. It is used as the learning agent for closed-loop control of morphing or highly reconfigurable systems. This paper develops a method whereby a state-space is adaptively discretized by progressively finer pseudorandom grids around the regions of interest within the state or learning space in an effort to break the Curse of Dimensionality. Utility of the method is demonstrated with application to the problem of a morphing airfoil, which is simulated by a computationally intensive computational fluid dynamics model. By setting the multiresolution method to define the region of interest by the goal the agent seeks, it is shown that this method with the pseudorandom grid can learn a specific goal within ±0.001 while reducing the total number of state-action pairs needed to achieve this level of specificity to less than 3000. © 2011 South China University of Technology, Academy of Mathematics and Systems Science, Chinese Academy of Sciences and Springer-Verlag Berlin Heidelberg.
引用
收藏
页码:431 / 439
页数:8
相关论文
共 34 条
  • [1] Multiresolution state-space discretization for Q-Learning with pseudorandomized discretization
    Amanda LAMPTON
    John VALASEK
    Mrinal KUMAR
    Journal of Control Theory and Applications, 2011, 9 (03) : 431 - 439
  • [2] Q-Learning Acceleration via State-space Partitioning
    Wei, Haoran
    Corder, Kevin
    Decker, Keith
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 293 - 298
  • [3] An adjustment method of the number of states on Q-learning segmenting state space adaptively
    Hamagami, T
    Hirata, H
    2003 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-5, CONFERENCE PROCEEDINGS, 2003, : 3062 - 3067
  • [4] An adjustment method of the number of states on Q-Learning segmenting state space adaptively
    Hamagami, Tornoki
    Koakutsu, Sefichi
    Hirata, Hironori
    ELECTRONICS AND COMMUNICATIONS IN JAPAN PART II-ELECTRONICS, 2007, 90 (09): : 75 - 86
  • [5] Labeling Q-learning in hidden state environments
    Hae-Yeon Lee
    Hiroyuki Kamaya
    Ken-ichi Abe
    Artificial Life and Robotics, 2002, 6 (4) : 181 - 184
  • [6] Intelligent Traffic Signal Duration Adaptation using Q-Learning with An Evolving State Space
    Gaikwad, Vinayak V.
    Kadarkar, Sanket S.
    Kasbekar, Gaurav S.
    2016 IEEE 84TH VEHICULAR TECHNOLOGY CONFERENCE (VTC FALL), 2016,
  • [7] Reinforcement distribution in continuous state action space fuzzy Q-learning: A novel approach
    Bonarini, A
    Montrone, F
    Restelli, M
    FUZZY LOGIC AND APPLICATIONS, 2006, 3849 : 40 - 45
  • [8] A Path Planning Algorithm for Space Manipulator Based on Q-Learning
    Li, Taiguo
    Li, Quanhong
    Li, Wenxi
    Xia, Jiagao
    Tang, Wenhua
    Wang, Weiwen
    PROCEEDINGS OF 2019 IEEE 8TH JOINT INTERNATIONAL INFORMATION TECHNOLOGY AND ARTIFICIAL INTELLIGENCE CONFERENCE (ITAIC 2019), 2019, : 1566 - 1571
  • [9] Virtual markets:: Q-learning sellers with simple state representation
    Akchurina, Natalia
    Buening, Hans Kleine
    AUTONOMOUS INTELLIGENT SYSTEMS: AGENTS AND DATA MINING, PROCEEDINGS, 2007, 4476 : 192 - +
  • [10] State-space segmentation for faster training reinforcement learning
    Kim, Jongrae
    IFAC PAPERSONLINE, 2022, 55 (25): : 235 - 240