A Model-based Factored Bayesian Reinforcement Learning Approach

被引:0
作者
Wu, Bo [1 ]
Feng, Yanpeng [1 ]
Zheng, Hongyan [1 ]
机构
[1] Shenzhen Polytech, Educ Technol & Informat Ctr, Shenzhen 518055, Peoples R China
来源
APPLIED SCIENCE, MATERIALS SCIENCE AND INFORMATION TECHNOLOGIES IN INDUSTRY | 2014年 / 513-517卷
关键词
MDPs; Bayesian; Reinforcement Learning;
D O I
10.4028/www.scientific.net/AMM.513-517.1092
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Bayesian reinforcement learning has turned out to be an effective solution to the optimal tradeoff between exploration and exploitation. However, in practical applications, the learning parameters with exponential growth are the main impediment for online planning and learning. To overcome this problem, we bring factored representations, model-based learning, and Bayesian reinforcement learning together in a new approach. Firstly, we exploit a factored representation to describe the states to reduce the size of learning parameters, and adopt Bayesian inference method to learn the unknown structure and parameters simultaneously. Then, we use an online point-based value iteration algorithm to plan and learn. The experimental results show that the proposed approach is an effective way for improving the learning efficiency in large-scale state spaces.
引用
收藏
页码:1092 / 1095
页数:4
相关论文
共 9 条
[1]  
DUFF M, 2002, THESIS U MASSASSACHU
[2]   Efficient solution algorithms for factored MDPs [J].
Guestrin, C ;
Koller, D ;
Parr, R ;
Venkataraman, S .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2003, 19 :399-468
[3]  
Poupart P., 2008, P ISAIM, P1025
[4]  
Poupart Pascal, 2006, ACM International Conference Proceeding Series, P697
[5]  
Ross S, 2011, J MACH LEARN RES, V12, P1729
[6]  
Ross Stephane, 2008, Uncertain Artif Intell, V2008, P476
[7]  
Sutton R.S., 2017, Introduction to reinforcement learning
[8]  
Wang Y., 2012, P ICML, P1135
[9]  
Wu B., 2013, RUANJIAN XUEBAO, V24, P25