Large-Scale Markov Decision Problems with KL Control Cost and its Application to Crowdsourcing

被引:0
|
作者
Abbasi-Yadkori, Yasin [1 ]
Bartlett, Peter L. [1 ,2 ]
Chen, Xi [3 ]
Malek, Alan [2 ]
机构
[1] Queensland Univ Technol, Brisbane, Qld 4001, Australia
[2] Univ Calif Berkeley, Berkeley, CA 94720 USA
[3] NYU, Stern Sch Business, New York, NY 10003 USA
基金
澳大利亚研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study average and total cost Markov decision problems with large state spaces. Since the computational and statistical cost of finding the optimal policy scales with the size of the state space, we focus on searching for near-optimality in a low-dimensional family of policies. In particular, we show that for problems with a Kullback-Leibler divergence cost function, we can recast policy optimization as a convex optimization and solve it approximately using a stochastic subgradient algorithm. This method scales in complexity with the family of policies but not the state space. We show that the performance of the resulting policy is close to the best in the low-dimensional family. We demonstrate the efficacy of our approach by optimizing a policy for budget allocation in crowd labeling, an important crowd-sourcing application.
引用
收藏
页码:1053 / 1062
页数:10
相关论文
共 50 条