On the effects of domain size and complexity in empirical distribution of reinforcement learning

被引:0
作者
Iwata, K [1 ]
Ikeda, K [1 ]
Sakai, H [1 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Dept Syst Sci, Kyoto 6068501, Japan
来源
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS | 2005年 / E88D卷 / 01期
关键词
reinforcement learning; Markov decision process; Lempel-Ziv coding; domain size; stochastic complexity;
D O I
10.1093/ietisy/E88-D.1.135
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We regard the events of a Markov decision process as the outputs from a Markov information source in order to analyze the randomness of an empirical sequence by the codeword length of the sequence. The randomness is an important viewpoint in reinforcement learning since the learning is to eliminate the randomness and to find an optimal policy. The occurrence of optimal empirical sequence also depends on the randomness. We then introduce the Lempel-Ziv coding for measuring the randomness which consists of the domain size and the stochastic complexity. In experimental results, we confirm that the learning and the occurrence of optimal empirical sequence depend on the randomness and show the fact that in early stages the randomness is mainly characterized by the domain size and as the number of time steps increases the randomness depends greatly on the complexity of Markov decision processes.
引用
收藏
页码:135 / 142
页数:8
相关论文
共 16 条