Efficient exploration through active learning for value function approximation in reinforcement learning

被引:15
作者
Akiyama, Takayuki [1 ]
Hachiya, Hirotaka [1 ]
Sugiyama, Masashi [1 ,2 ]
机构
[1] Tokyo Inst Technol, Dept Comp Sci, Meguro Ku, Tokyo 1528552, Japan
[2] Japan Sci & Technol Agcy, PRESTO, Tokyo, Japan
关键词
Reinforcement learning; Markov decision process; Least-squares policy iteration; Active learning; Batting robot; REGRESSION;
D O I
10.1016/j.neunet.2009.12.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Appropriately designing sampling policies is highly important for obtaining better control policies in reinforcement learning. In this paper, we first show that the least-squares policy iteration (LSPI) framework allows us to employ statistical active learning methods for linear regression. Then we propose a design method of good sampling policies for efficient exploration, which is particularly useful when the sampling cost of immediate rewards is high. The effectiveness of the proposed method, which we call active policy iteration (API), is demonstrated through simulations with a batting robot. (C) 2010 Elsevier Ltd. All rights reserved.
引用
收藏
页码:639 / 648
页数:10
相关论文
共 22 条