State Abstraction in Reinforcement Learning by Eliminating Useless Dimensions

被引:1
作者
Cheng, Zhao [1 ]
Ray, Laura E. [1 ]
机构
[1] Dartmouth Coll, Thayer Sch Engn, Hanover, NH 03755 USA
来源
2014 13TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA) | 2014年
关键词
reinforcement learning; intelligent agent; state abstraction; complexity reduction;
D O I
10.1109/ICMLA.2014.22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Q-learning and other linear dynamic learning algorithms are subject to Bellman's curse of dimensionality for any realistic learning problem. This paper introduces a framework for satisficing state abstraction-one that reduces state dimensionality, improving convergence and reducing computational and memory resources-by eliminating useless state dimensions. Statistical parameters that are dependent on the state and Q-values identify the relevance of a given state space to a task space and allow state elements that contribute least to task learning to be discarded. Empirical results of applying state abstraction to a canonical single-agent path planning task and to a more difficult multi-agent foraging problem demonstrate utility of the proposed methods in improving learning convergence and performance in resource-constrained learning problems.
引用
收藏
页码:105 / 110
页数:6
相关论文
共 18 条
[1]  
ANDRE D, 2002, P 18 NAT C ART INT
[2]  
[Anonymous], 2002, SURVEY DIMENSION RED
[3]  
[Anonymous], 1995, THESIS
[4]  
Asadi M., 2005, ARTIFICIAL INTELLIGE
[5]  
Asadi M., 2004, Proceedings of the Seventeenth International FLAIRS Conference, P509
[6]  
Bertsekas Dimitri P., 2018, Abstract Dynamic Programming, V2nd
[7]  
Boyan J. A., 1995, Advances in Neural Information Processing Systems 7, P369
[8]   AUTOMATIC COMPLEXITY REDUCTION IN REINFORCEMENT LEARNING [J].
Chiu, Chung-Cheng ;
Soo, Von-Wun .
COMPUTATIONAL INTELLIGENCE, 2010, 26 (01) :1-25
[9]  
Guyon I., 2003, Journal of Machine Learning Research, V3, P1157, DOI 10.1162/153244303322753616
[10]  
HENGST B, 2003, 0308 UNSW CSE TR