Extracting finite-state representations from recurrent neural networks trained on chaotic symbolic sequences

被引:30
作者
Tino, P [1 ]
Köteles, M [1 ]
机构
[1] Slovak Tech Univ Bratislava, Dept Comp Sci & Engn, Bratislava 81219, Slovakia
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1999年 / 10卷 / 02期
基金
奥地利科学基金会;
关键词
chaotic symbolic sequences; entropy spectra; fractal geometry; recurrent neural network; stochastic automata;
D O I
10.1109/72.750555
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While much work has been done in neural-based modeling of real-valued chaotic time series, little effort has been devoted to address similar problems in the symbolic domain. We investigate the knowledge induction process associated With training recurrent neural net work (RNN's) on single long chaotic symbolic sequences, El en though training RNN's to predict the nest symbol leaves the standard performance measures such as the mean square error on the network output virtually unchanged, the network nevertheless do extract a lot of knowledge. We monitor the knowledge extraction process by considering the networks stochastic sources and letting them generate sequences which are then confronted with the training sequence via information theoretic entropy and cross-entropy measures, We also study the possibility of reformulating the knowledge gained by RNN's in a compact and easy-to-analyze form of finite-state stochastic machines. The experiments are performed on two sequences with different "complexities" measured by the size and state transition structure of the induced Crutchfield's epsilon-machines, We find that, with respect to the original RNN's, the extracted machines can achieve comparable or even better entropy and cross-entropy performance. Moreover. RNN's reflect the training sequence complexity in their dynamical state representations that can in turn be reformulated using finite-state means. Our findings are confirmed by a much more detailed analysis of model generated sequences through the statistical mechanical metaphor of entropy spectra, We also introduce a visual representation of allowed block structure in the studied sequences that, besides having nice theoretical properties, allows on the topological level for an illustrative insight into both RNN training and finite-state stochastic machine extraction processes.
引用
收藏
页码:284 / 302
页数:19
相关论文
共 45 条
[1]  
Anderberg M.R., 1973, Probability and Mathematical Statistics
[2]  
[Anonymous], MACHINE LEARNING
[3]  
[Anonymous], 1991, ACM Computing Surveys
[4]  
[Anonymous], 1957, MATH FDN INFORM THEO
[5]  
[Anonymous], 1995, THERMODYNAMICS CHAOT
[6]  
Barnsley M. F., 2014, Fractals Everywhere
[7]  
BOX GEP, 1970, TIME SERIES ANAL FOR
[8]   DYNAMIC CELL STRUCTURE LEARNS PERFECTLY TOPOLOGY PRESERVING MAP [J].
BRUSKE, J ;
SOMMER, G .
NEURAL COMPUTATION, 1995, 7 (04) :845-865
[9]   The dynamics of discrete-time computation, with application to recurrent neural networks and finite state machine extraction [J].
Casey, M .
NEURAL COMPUTATION, 1996, 8 (06) :1135-1178
[10]   Finite State Automata and Simple Recurrent Networks [J].
Cleeremans, Axel ;
Servan-Schreiber, David ;
McClelland, James L. .
NEURAL COMPUTATION, 1989, 1 (03) :372-381