Predictive feature selection for genetic policy search

被引:3
作者
Loscalzo, Steven [1 ]
Wright, Robert [2 ]
Yu, Lei [2 ]
机构
[1] AFRL Informat Directorate, Rome, NY 13441 USA
[2] SUNY Binghamton, Binghamton, NY 13902 USA
关键词
Genetic policy search; Feature selection; Dimensionality reduction; Reinforcement learning; REINFORCEMENT; CLASSIFICATION; EXPLORATION; STATE;
D O I
10.1007/s10458-014-9268-y
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automatic learning of control policies is becoming increasingly important to allow autonomous agents to operate alongside, or in place of, humans in dangerous and fast-paced situations. Reinforcement learning (RL), including genetic policy search algorithms, comprise a promising technology area capable of learning such control policies. Unfortunately, RL techniques can take prohibitively long to learn a sufficiently good control policy in environments described by many sensors (features). We argue that in many cases only a subset of available features are needed to learn the task at hand, since others may represent irrelevant or redundant information. In this work, we propose a predictive feature selection framework that analyzes data obtained during execution of a genetic policy search algorithm to identify relevant features on-line. This serves to constrain the policy search space and reduces the time needed to locate a sufficiently good policy by embedding feature selection into the process of learning a control policy. We explore this framework through an instantiation called predictive feature selection embedded in neuroevolution of augmenting topology (NEAT), or PFS-NEAT. In an empirical study, we demonstrate that PFS-NEAT is capable of enabling NEAT to successfully find good control policies in two benchmark environments, and show that it can outperform three competing feature selection algorithms, FS-NEAT, FD-NEAT, and SAFS-NEAT, in several variants of these environments.
引用
收藏
页码:754 / 786
页数:33
相关论文
共 57 条
[1]  
[Anonymous], 2010, PRINCIPAL COMPONENT
[2]  
[Anonymous], 1963, DYNAMIC PROGRAMMING
[3]  
[Anonymous], 2009, Proceedings of the 26th International Conference on Machine Learning
[4]  
[Anonymous], 1998, Reinforcement Learning: An Introduction
[5]  
[Anonymous], 2011, Approximate Dynamic Programming: Solving the Curses of Dimensionality
[6]   A survey of robot learning from demonstration [J].
Argall, Brenna D. ;
Chernova, Sonia ;
Veloso, Manuela ;
Browning, Brett .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2009, 57 (05) :469-483
[7]  
Bengio Y., 2009, P 26 ANN INT C MACHI, P41, DOI DOI 10.1145/1553374.1553380
[8]  
Bohm Niko, 2004, LWA 2004 LERNEN WISS, P118
[9]  
Boutilier C, 1999, J ARTIF INTELL RES, V11, P1
[10]  
Cannady James., 2000, Proceedings of the 23rd national information systems security conference, P1