首页 | 本学科首页   官方微博 | 高级检索  
     


FERNN: An Algorithm for Fast Extraction of Rules from Neural Networks
Authors:Rudy Setiono  Wee Kheng Leow
Affiliation:(1) School of Computing, National University of Singapore, Lower Kent Ridge Road, Singapore, 119260;(2) School of Computing, National University of Singapore, Lower Kent Ridge Road, Singapore, 119260
Abstract:Before symbolic rules are extracted from a trained neural network, the network is usually pruned so as to obtain more concise rules. Typical pruning algorithms require retraining the network which incurs additional cost. This paper presents FERNN, a fast method for extracting rules from trained neural networks without network retraining. Given a fully connected trained feedforward network with a single hidden layer, FERNN first identifies the relevant hidden units by computing their information gains. For each relevant hidden unit, its activation values is divided into two subintervals such that the information gain is maximized. FERNN finds the set of relevant network connections from the input units to this hidden unit by checking the magnitudes of their weights. The connections with large weights are identified as relevant. Finally, FERNN generates rules that distinguish the two subintervals of the hidden activation values in terms of the network inputs. Experimental results show that the size and the predictive accuracy of the tree generated are comparable to those extracted by another method which prunes and retrains the network.
Keywords:rule extraction  penalty function  MofN rule  DNF rule  decision tree
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号