首页 | 本学科首页   官方微博 | 高级检索  
     


Soft Margins for AdaBoost
Authors:G Rätsch  T Onoda  K-R Müller
Affiliation:(1) GMD FIRST, Kekuléstr. 7, 12489 Berlin, Germany;(2) CRIEPI, Komae-shi, 2-11-1 Iwado Kita, Tokyo, Japan;(3) GMD FIRST, Kekuléstr. 7, 12489 Berlin, Germany;(4) University of Potsdam, Neues Palais 10, 14469 Potsdam, Germany
Abstract:Recently ensemble methods like ADABOOST have been applied successfully in many problems, while seemingly defying the problems of overfitting.ADABOOST rarely overfits in the low noise regime, however, we show that it clearly does so for higher noise levels. Central to the understanding of this fact is the margin distribution. ADABOOST can be viewed as a constraint gradient descent in an error function with respect to the margin. We find that ADABOOST asymptotically achieves a hard margin distribution, i.e. the algorithm concentrates its resources on a few hard-to-learn patterns that are interestingly very similar to Support Vectors. A hard margin is clearly a sub-optimal strategy in the noisy case, and regularization, in our case a ldquomistrustrdquo in the data, must be introduced in the algorithm to alleviate the distortions that single difficult patterns (e.g. outliers) can cause to the margin distribution. We propose several regularization methods and generalizations of the original ADABOOST algorithm to achieve a soft margin. In particular we suggest (1) regularized ADABOOSTREG where the gradient decent is done directly with respect to the soft margin and (2) regularized linear and quadratic programming (LP/QP-) ADABOOST, where the soft margin is attained by introducing slack variables.Extensive simulations demonstrate that the proposed regularized ADABOOST-type algorithms are useful and yield competitive results for noisy data.
Keywords:ADABOOST  arcing  large margin  soft margin  classification  support vectors
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号