Neural Estimator of Information for Time-Series Data with Dependency

被引:3
作者
Molavipour, Sina [1 ]
Ghourchian, Hamid [1 ]
Bassi, German [2 ]
Skoglund, Mikael [1 ]
机构
[1] KTH Royal Inst Technol, Sch Elect Engn & Comp Sci EECS, S-10044 Stockholm, Sweden
[2] Ericsson Res, S-16483 Stockholm, Sweden
基金
瑞典研究理事会;
关键词
neural networks; conditional mutual information; directed information; Markov source; variational bound; DIRECTED INFORMATION; MUTUAL INFORMATION; TRANSFER ENTROPY; CONNECTIVITY; FEEDFORWARD; NETWORKS;
D O I
10.3390/e23060641
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Novel approaches to estimate information measures using neural networks are well-celebrated in recent years both in the information theory and machine learning communities. These neural-based estimators are shown to converge to the true values when estimating mutual information and conditional mutual information using independent samples. However, if the samples in the dataset are not independent, the consistency of these estimators requires further investigation. This is of particular interest for a more complex measure such as the directed information, which is pivotal in characterizing causality and is meaningful over time-dependent variables. The extension of the convergence proof for such cases is not trivial and demands further assumptions on the data. In this paper, we show that our neural estimator for conditional mutual information is consistent when the dataset is generated with samples of a stationary and ergodic source. In other words, we show that our information estimator using neural networks converges asymptotically to the true value with probability one. Besides universal functional approximation of neural networks, a core lemma to show the convergence is Birkhoff's ergodic theorem. Additionally, we use the technique to estimate directed information and demonstrate the effectiveness of our approach in simulations.
引用
收藏
页数:28
相关论文
共 49 条
[1]  
Aharoni Z., ARXIV200304179
[2]  
[Anonymous], 1990, P INT S INF THEOR AP
[3]  
[Anonymous], 2007, 2 COURSE PROBABILITY
[4]  
[Anonymous], 2014, UNDERSTANDING MACHIN
[5]  
Belghazi I., 2018, P 35 INT C MACH LEAR, V80, P531
[6]   THE INDIVIDUAL ERGODIC THEOREM OF INFORMATION-THEORY [J].
BREIMAN, L .
ANNALS OF MATHEMATICAL STATISTICS, 1957, 28 (03) :809-811
[7]   Inferring neuronal network functional connectivity with directed information [J].
Cai, Zhiting ;
Neveu, Curtis L. ;
Baxter, Douglas A. ;
Byrne, John H. ;
Aazhang, Behnaam .
JOURNAL OF NEUROPHYSIOLOGY, 2017, 118 (02) :1055-1069
[8]   Statistical assessment of nonlinear causality:: application to epileptic EEG signals [J].
Chávez, M ;
Martinerie, J ;
Le Van Quyen, M .
JOURNAL OF NEUROSCIENCE METHODS, 2003, 124 (02) :113-128
[9]  
Collomb G., 1985, Statistics, V16, P297, DOI [10.1080/02331888508801858, DOI 10.1080/02331888508801858]
[10]  
Cover T. M., 2006, Elements of Information Theory: Solutions to Problems, V2nd