Memory Complexity of Estimating Entropy and Mutual Information

被引:0
作者
Berg, Tomer [1 ]
Ordentlich, Or [2 ]
Shayevitz, Ofer [1 ]
机构
[1] Tel Aviv Univ, Dept Elect Engn Syst, IL-69978 Tel Aviv, Israel
[2] Hebrew Univ Jerusalem, Sch Comp Sci & Engn, IL-91904 Jerusalem, Israel
基金
以色列科学基金会;
关键词
Entropy; Estimation; Complexity theory; Memory management; Mutual information; Lower bound; Additives; Upper bound; Testing; Automata; Memory complexity; sample complexity; entropy estimation; mutual information estimation; finite memory algorithms;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We observe an infinite sequence of independent identically distributed random variables X-1, X-2,& mldr; drawn from an unknown distribution p over [n] , and our goal is to estimate the entropy H(p) = -E[logp(X)] within an epsilon -additive error. To that end, at each time point we are allowed to update a finite-state machine with S states, using a possibly randomized but time-invariant rule, where each state of the machine is assigned an entropy estimate. Our goal is to characterize the minimax memory complexity S & lowast; of this problem, which is the minimal number of states for which the estimation task is feasible with probability at least 1 - delta asymptotically, uniformly in p. Specifically, we show that there exist universal constants C-1 and C-2 such that S* <= C-1 & sdot;n(log n)(4)/epsilon(2)delta for epsilon not too small, and S* >= C-2 & sdot;max{n, log n/epsilon} for epsilon not too large. The upper bound is proved using approximate counting to estimate the logarithm of p, and a finite memory bias estimation machine to estimate the expectation operation. The lower bound is proved via a reduction of entropy estimation to uniformity testing. We also apply these results to derive bounds on the memory complexity of mutual information estimation.
引用
收藏
页码:3334 / 3349
页数:16
相关论文
共 37 条
[1]  
Acharya J., 2019, arXiv
[2]  
Aliakbarpour M, 2022, Arxiv, DOI arXiv:2205.09804
[3]  
[Anonymous], 2015, P C LEARN THEOR JUN
[4]   Convergence properties of functional estimates for discrete distributions [J].
Antos, A ;
Kontoyiannis, I .
RANDOM STRUCTURES & ALGORITHMS, 2001, 19 (3-4) :163-193
[5]  
Basharin G. P., 1959, Theory Probab. Its Appl, V4, P333
[6]  
Berg T, 2022, PR MACH LEARN RES, V178
[7]   Statistical Inference With Limited Memory: A Survey [J].
Berg, Tomer ;
Ordentlich, Or ;
Shayevitz, Ofer .
IEEE JOURNAL ON SELECTED AREAS IN INFORMATION THEORY, 2024, 5 :623-644
[8]  
Berg T, 2021, PR MACH LEARN RES, V134, P566
[9]   Estimating Entropy and Entropy Norm on Data Streams [J].
Chakrabarti, Amit ;
Do Ba, Khanh ;
Muthukrishnan, S. .
INTERNET MATHEMATICS, 2006, 3 (01) :63-78
[10]  
Chien S., 2010, P ICS JAN, P251