FERNN: An Algorithm for Fast Extraction of Rules from Neural Networks |
| |
Authors: | Rudy Setiono Wee Kheng Leow |
| |
Affiliation: | (1) School of Computing, National University of Singapore, Lower Kent Ridge Road, Singapore, 119260;(2) School of Computing, National University of Singapore, Lower Kent Ridge Road, Singapore, 119260 |
| |
Abstract: | Before symbolic rules are extracted from a trained neural network, the network is usually pruned so as to obtain more concise rules. Typical pruning algorithms require retraining the network which incurs additional cost. This paper presents FERNN, a fast method for extracting rules from trained neural networks without network retraining. Given a fully connected trained feedforward network with a single hidden layer, FERNN first identifies the relevant hidden units by computing their information gains. For each relevant hidden unit, its activation values is divided into two subintervals such that the information gain is maximized. FERNN finds the set of relevant network connections from the input units to this hidden unit by checking the magnitudes of their weights. The connections with large weights are identified as relevant. Finally, FERNN generates rules that distinguish the two subintervals of the hidden activation values in terms of the network inputs. Experimental results show that the size and the predictive accuracy of the tree generated are comparable to those extracted by another method which prunes and retrains the network. |
| |
Keywords: | rule extraction penalty function MofN rule DNF rule decision tree |
本文献已被 SpringerLink 等数据库收录! |
|