Shaping Bayesian Network Based Reinforcement Learning

被引:0
作者
Song, Jiong [1 ]
Jin, Zhao [2 ]
机构
[1] Yunnan Jiao Tong Vocat & Tech Coll, Kunming 650101, Peoples R China
[2] Yunnan Univ, Kunming, Peoples R China
来源
2012 INTERNATIONAL CONFERENCE ON INDUSTRIAL CONTROL AND ELECTRONICS ENGINEERING (ICICEE) | 2012年
关键词
component; Reinforcement learning; State transition; Shaping Bayesian network; Accelerate learning;
D O I
10.1109/ICICEE.2012.199
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The trial-and-error mechanism of reinforcement learning essentially is a kind of exhaustive search, which is also the major reason to cause reinforcement learning being slow and time-consuming. We present an approach to model the state transitions in agent' s exploration by Shaping Bayesian Network, which can be used to shape agent for bias exploration towards the most promising regions of state space and thereby reduce exploration and accelerate learning. The experiment results show this approach can significantly improve agent's performance and shorten learning time. More importantly, this approach provides a kind of way to make agent can take advantage of its own experience to accelerate learning.
引用
收藏
页码:742 / 745
页数:4
相关论文
共 9 条
[1]  
Amizadeh S., 2007, 2007 INT C ADV INT M, P1
[2]  
[Anonymous], 2008, P 25 INT C MACHINE L
[3]  
Doshi Finale, 2008, Proc Int Conf Mach Learn, V301, P256
[4]   Improving reinforcement learning by using sequence trees [J].
Girgin, Sertan ;
Polat, Faruk ;
Alhajj, Reda .
MACHINE LEARNING, 2010, 81 (03) :283-331
[5]  
Jin Z, 2010, LECT NOTES COMPUT SC, V6377, P49
[6]  
Pearl J., 1988, PROBABILISTIC REASON, P79
[7]  
Price Bob., 2003, P 18 INT JOINT C ART, P712
[8]  
Sutton R.S., 2017, Introduction to reinforcement learning
[9]  
Wang Tao, 2005, Proceedings of the 22nd international conference on Machine learning (ICML), P956