Minimax redundancy for the class of memoryless sources

被引:68
|
作者
Xie, Q
Barron, AR
机构
[1] Department of Statistics, Yale University, New Haven
基金
美国国家科学基金会;
关键词
universal noiseless coding; minimax redundancy; minimax total relative entropy risk; Jeffreys' prior; asymptotic least favorable prior;
D O I
10.1109/18.556120
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Let X(n) = (X(1), ..., X(n)) be a memoryless source with unknown distribution on a finite alphabet of size k. We identify the asymptotic minimax coding redundancy for this class of sources, and provide a sequence of asymptotically mini max codes, Equivalently, we determine the limiting behavior of the minimax relative entropy min(QXn) max(PXn) D(P-Xn\\Q(Xn)), where the maximum is over all independent and identically distributed (i.i.d.) source distributions and the minimum is over all joint distributions, We show in this paper that the minimax redundancy minus ((k - 1)/2) log (n/(2 pi e)) converges to log integral root det I(theta) d theta = log (Gamma(1/2)(k)/Gamma(k/2)), where I(theta) is the Fisher information and the integral is over the whole probability simplex, The Bayes strategy using Jeffreys' prior is shown to be asymptotically maximin but not asymptotically minimax in our setting, The boundary risk using Jeffreys' prior is higher than that of interior points, We provide a sequence of modifications of Jeffreys' prior that put some prior mass near the boundaries of the probability simplex to pull down that risk to the asymptotic minimax level in the limit.
引用
收藏
页码:646 / 657
页数:12
相关论文
共 7 条