Extraction of rules from artificial neural networks for nonlinear regression

被引:126
作者
Setiono, R [1 ]
Leow, WK
Zurada, JM
机构
[1] Natl Univ Singapore, Sch Comp, Singapore 119260, Singapore
[2] Univ Louisville, Dept Elect & Comp Engn, Louisville, KY 40208 USA
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2002年 / 13卷 / 03期
关键词
network pruning; regression; rule extraction;
D O I
10.1109/TNN.2002.1000125
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks (NNs) have been successfully applied to solve a variety of application problems including classification and function approximation. They are especially useful as function approximators because they do not require prior knowledge of the input data distribution and they have been shown to be universal approximators. In many applications, it is desirable to extract knowledge that can explain how the problems are solved by the networks. Most existing approaches have focused on extracting symbolic rules for classification. Few methods have been devised to extract rules from trained NNs for regression. This article presents an approach for extracting rules from trained NNs for regression. Each rule in the extracted rule set corresponds to a subregion of the input space and a linear function involving the relevant input attributes of the data approximates the network output for all data samples in this subregion. Extensive experimental results on 32 benchmark data sets demonstrate the effectiveness of the proposed approach in generating accurate regression rules.
引用
收藏
页码:564 / 577
页数:14
相关论文
共 36 条
[1]   Survey and critique of techniques for extracting rules from trained artificial neural networks [J].
Andrews, R ;
Diederich, J ;
Tickle, AB .
KNOWLEDGE-BASED SYSTEMS, 1995, 8 (06) :373-389
[2]  
[Anonymous], 1992, Proceedings of the 5th Australian Joint Conference on Artificial Intelligence (AI'92), DOI DOI 10.1142/9789814536271
[3]  
Ash T., 1989, Connection Science, V1, P365, DOI 10.1080/09540098908915647
[4]   DETERMINING INPUT FEATURES FOR MULTILAYER PERCEPTRONS [J].
BELUE, LM ;
BAUER, KW .
NEUROCOMPUTING, 1995, 7 (02) :111-121
[5]  
BLASSIG R, 1994, ADV NEURAL INFORMATI, V6, P1093
[6]  
Dennis, 1996, NUMERICAL METHODS UN
[7]   Naive Bayes for regression [J].
Frank, E ;
Trigg, L ;
Holmes, G ;
Witten, IH .
MACHINE LEARNING, 2000, 41 (01) :5-25
[8]   Function approximation with spiked random networks [J].
Gelenbe, E ;
Mao, ZH ;
Li, YD .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1999, 10 (01) :3-9
[9]  
Gill M., 1981, Practical Optimization
[10]   Generalized analytic rule extraction for feedforward neural networks [J].
Gupta, A ;
Park, S ;
Lam, SM .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 1999, 11 (06) :985-991