Rule extraction from recurrent neural networks: A taxonomy and review

被引:101
作者
Jacobsson, H [1 ]
机构
[1] Univ Skovde, Sch Humanities & Informat, Skovde, Sweden
[2] Univ Sheffield, Dept Comp Sci, Sheffield S10 2TN, S Yorkshire, England
关键词
D O I
10.1162/0899766053630350
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Rule extraction (RE) from recurrent neural networks (RNNs) refers to finding models of the underlying RNN, typically in the form of finite state machines, that mimic the network to a satisfactory degree while having the advantage of being more transparent. RE from RNNs can be argued to allow a deeper and more profound form of analysis of RNNs than other, more or less ad hoc methods. RE may give us understanding of RNNs in the intermediate levels between quite abstract theoretical knowledge of RNNs as a class of computing devices and quantitative performance evaluations of RNN instantiations. The development of techniques for extraction of rules from RNNs has been an active field since the early 1990s. This article reviews the progress of this development and analyzes it in detail. In order to structure the survey and evaluate the techniques, a taxonomy specifically designed for this purpose has been developed. Moreover, important open research issues are identified that, if addressed properly, possibly can give the field a significant push forward.
引用
收藏
页码:1223 / 1263
页数:41
相关论文
共 124 条
[1]  
ALQUEZAR R, 1994, ADV PATTERN RECOGNIT, P48
[2]  
ALQUEZAR R, 1994, P ICGI 94 BERL, P203
[3]  
ALQUEZAR R, 1997, 7 S NAC REC FORM AN, V1, P49
[4]   Survey and critique of techniques for extracting rules from trained artificial neural networks [J].
Andrews, R ;
Diederich, J ;
Tickle, AB .
KNOWLEDGE-BASED SYSTEMS, 1995, 8 (06) :373-389
[5]  
[Anonymous], PROCEEDINGS OF THE S
[6]  
[Anonymous], P IEEE IJCNN SEATTL
[7]  
Bakker B, 2000, FROM ANIM ANIMAT, P51
[8]  
BAKKER B, 2004, THESIS LEIDEN U
[9]   A taxonomy for spatiotemporal connectionist networks revisited:: The unsupervised case [J].
Barreto, GD ;
Araújo, AFR .
NEURAL COMPUTATION, 2003, 15 (06) :1255-1320
[10]   LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT [J].
BENGIO, Y ;
SIMARD, P ;
FRASCONI, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02) :157-166