On the Accurate Estimation of Information-Theoretic Quantities from Multi-Dimensional Sample Data

被引:4
作者
Chaves, Manuel Alvarez [1 ]
Gupta, Hoshin V. [2 ]
Ehret, Uwe [3 ]
Guthke, Anneli [1 ]
机构
[1] Univ Stuttgart, Stuttgart Ctr Simulat Sci, Cluster Excellence EXC 2075, D-70569 Stuttgart, Germany
[2] Univ Arizona, Hydrol & Atmospher Sci, Tucson, AZ 85721 USA
[3] Inst Water & River Basin Management, Karlsruhe Inst Technol KIT, D-76131 Karlsruhe, Germany
关键词
information theory; non-parametric estimation; entropy; mutual information; Kullback-Leibler divergence; relative entropy; data; binning; kernel density estimation; k-nearest neighbors; k-NN; ENTROPY; HISTOGRAM; CHOICE;
D O I
10.3390/e26050387
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Using information-theoretic quantities in practical applications with continuous data is often hindered by the fact that probability density functions need to be estimated in higher dimensions, which can become unreliable or even computationally unfeasible. To make these useful quantities more accessible, alternative approaches such as binned frequencies using histograms and k-nearest neighbors (k-NN) have been proposed. However, a systematic comparison of the applicability of these methods has been lacking. We wish to fill this gap by comparing kernel-density-based estimation (KDE) with these two alternatives in carefully designed synthetic test cases. Specifically, we wish to estimate the information-theoretic quantities: entropy, Kullback-Leibler divergence, and mutual information, from sample data. As a reference, the results are compared to closed-form solutions or numerical integrals. We generate samples from distributions of various shapes in dimensions ranging from one to ten. We evaluate the estimators' performance as a function of sample size, distribution characteristics, and chosen hyperparameters. We further compare the required computation time and specific implementation challenges. Notably, k-NN estimation tends to outperform other methods, considering algorithmic implementation, computational efficiency, and estimation accuracy, especially with sufficient data. This study provides valuable insights into the strengths and limitations of the different estimation methods for information-theoretic quantities. It also highlights the significance of considering the characteristics of the data, as well as the targeted information-theoretic quantity when selecting an appropriate estimation technique. These findings will assist scientists and practitioners in choosing the most suitable method, considering their specific application and available data. We have collected the compared estimation methods in a ready-to-use open-source Python 3 toolbox and, thereby, hope to promote the use of information-theoretic quantities by researchers and practitioners to evaluate the information in data and models in various disciplines.
引用
收藏
页数:34
相关论文
共 50 条
[1]   NONPARAMETRIC ESTIMATION OF ENTROPY FOR ABSOLUTELY CONTINUOUS DISTRIBUTIONS [J].
AHMAD, IA ;
LIN, PE .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1976, 22 (03) :372-375
[2]  
Ao ZQ, 2022, Proceedings of the AAAI Conference on Artificial Intelligence, V36, P9990, DOI [10.1609/aaai.v36i9.21237, 10.1609/aaai.v36i9.21237]
[3]   Shannon Entropy and Mutual Information for Multivariate Skew-Elliptical Distributions [J].
Arellano-Valle, Reinaldo B. ;
Contreras-Reyes, Javier E. ;
Genton, Marc G. .
SCANDINAVIAN JOURNAL OF STATISTICS, 2013, 40 (01) :42-62
[4]  
Beirlant Jan., 1997, INT J MATH STAT SCI, V6, P17
[5]  
Belghazi MI, 2018, PR MACH LEARN RES, V80
[6]  
Bishop Christopher M., 2006, Pattern recognition and machine learning
[7]  
Bossomaier T., 2016, INTRO TRANSFER ENTRO, P65, DOI DOI 10.1007/978-3-319-43222-9
[8]  
Cover T.M., 2006, Elements of information theory. Wiley series in telecommunications and signal processing
[9]   Entropy expressions for multivariate continuous distributions [J].
Darbellay, GA ;
Vajda, I .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2000, 46 (02) :709-712
[10]   On the Kozachenko-Leonenko entropy estimator [J].
Delattre, Sylvain ;
Fournier, Nicolas .
JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2017, 185 :69-93