首页 | 本学科首页   官方微博 | 高级检索  
     


Combining Different Methods and Numbers of Weak Decision Trees
Authors:Patrice Latinne  Olivier Debeir  Christine Decaestecker
Affiliation:(1) IRIDIA (Artificial Intelligence Department), Université Libre de Bruxelles, Brussels, Belgium, BE;(2) Information and Decision Systems, Université Libre de Bruxelles, Brussels, Belgium, BE;(3) Laboratory of Histopathology, Université Libre de Bruxelles, Brussels, Belgium, BE
Abstract:Several ways of manipulating a training set have shown that weakened classifier combination can improve prediction accuracy. In the present paper, we focus on learning set sampling (Breiman’s Bagging) and random feature subset selections (Ho’s Random Subspaces). We present a combination scheme labelled ‘Bagfs’, in which new learning sets are generated on the basis of both bootstrap replicates and random subspaces. The performances of the three methods (Bagging, Random Subspaces and Bagfs) are compared to the standard Adaboost algorithm. All four methods are assessed by means of a decision-tree inducer (C4.5). In addition, we also study whether the number and the way in which they are created has a significant influence on the performance of their combination. To answer these two questions, we undertook the application of the McNemar test of significance and the Kappa degree-of-agreement. The results, obtained on 23 conventional databases, show that on average, Bagfs exhibits the best agreement between prediction and supervision. Received: 17 November 2000, Received in revised form: 30 October 2001, Accepted: 13 December 2001
Keywords:: Bagging   Boosting   Decision trees   Ensemble learning   Random subspaces
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号