Minimum interpretation by autoencoder-based serial and enhanced mutual information production

被引:0
作者
Ryotaro Kamimura
机构
[1] Tokai University,IT Education Center
来源
Applied Intelligence | 2020年 / 50卷
关键词
AutoEncoder; Minimum interpretation; Enhanced information; Generalization; Interpretation; Mutual information;
D O I
暂无
中图分类号
学科分类号
摘要
The present paper aims to propose an information-theoretic method for interpreting the inference mechanism of neural networks. The new method aims to interpret the inference mechanism minimally by disentangling complex information into simpler and easily interpretable information. This disentanglement of complex information can be realized by maximizing mutual information between input patterns and the corresponding neurons. However, because the use of mutual information has faced difficulty in computation, we use the well-known autoencoder to increase mutual information by re-interpreting the sparsity constraint, which is considered a device to increase mutual information. The computational procedures to increase mutual information are decomposed into the serial operation of equal use of neurons and specific responses to input patterns. The specific responses are realized by enhancing the results by the equal use of neurons. The method was applied to three data sets: the glass, office equipment, and pulsar data sets. With all three data sets, we could observe that, when the number of neurons was forced to increase, mutual information could be increased. Then, collective weights, or average collectively treated weights, showed that the method could extract the simple and linear relations between inputs and targets, making it possible to interpret the inference mechanism minimally.
引用
收藏
页码:2423 / 2448
页数:25
相关论文
共 70 条
[1]  
Linsker R(1988)Self-organization in a perceptual network Computer 21 105-117
[2]  
Linsker R(1989)How to generate ordered maps by maximizing the mutual information between input and output signals Neural Comput 1 402-411
[3]  
Linsker R(1992)Local synaptic learning rules suffice to maximize mutual information in a linear network Neural Comput 4 691-702
[4]  
Linsker R(2005)Improved local learning rule for information maximization and related applications Neural Netw 18 261-265
[5]  
Becker S(1996)Mutual information maximization: models of cortical self-organization Netw Comput Neural Syst 7 7-31
[6]  
Deco G(1995)Unsupervised mutual information criterion for elimination of overtraining in supervised multilayer networks Neural Comput 7 86-107
[7]  
Finnoff W(2000)Information theoretic learning Unsupervised Adaptive Filtering 1 265-319
[8]  
Zimmermann H(2003)Feature extraction by non-parametric mutual information maximization J Mach Learn Res 3 1415-1438
[9]  
Principe JC(2004)Fast binary feature selection with conditional mutual information J Mach Learn Res 5 1531-1555
[10]  
Xu D(2005)Estimating optimal feature subsets using efficient estimation of high-dimensional mutual information IEEE Trans Neural Netw 16 213-224