首页 | 本学科首页   官方微博 | 高级检索  
     

基于指数损失和0-1损失的在线Boosting算法
引用本文:侯杰,茅耀斌,孙金生.基于指数损失和0-1损失的在线Boosting算法[J].自动化学报,2014,40(4):635-642.
作者姓名:侯杰  茅耀斌  孙金生
作者单位:1.南京理工大学自动化学院 南京 210094
基金项目:国家自然科学基金(60974129)资助
摘    要:推导了使用指数损失函数和0-1损失函数的Boosting 算法的严格在线形式,证明这两种在线Boosting算法最大化样本间隔期望、最小化样本间隔方差.通过增量估计样本间隔的期望和方差,Boosting算法可应用于在线学习问题而不损失分类准确性. UCI数据集上的实验表明,指数损失在线Boosting算法的分类准确性与批量自适应 Boosting (AdaBoost)算法接近,远优于传统的在线Boosting;0-1损失在线Boosting算法分别最小化正负样本误差,适用于不平衡数据问题,并且在噪声数据上分类性能更为稳定.

关 键 词:AdaBoost    在线学习    特征选择    不平衡数据
收稿时间:2013-06-05

Online Boosting Algorithms Based on Exponential and 0-1 Loss
HOU Jie,MAO Yao-Bin,SUN Jin-Sheng.Online Boosting Algorithms Based on Exponential and 0-1 Loss[J].Acta Automatica Sinica,2014,40(4):635-642.
Authors:HOU Jie  MAO Yao-Bin  SUN Jin-Sheng
Affiliation:1.Institute of Automation, Nanjing University of Science and Technology, Nanjing 210094
Abstract:In this paper, strict derivation for the online form of Boosting algorithms using exponential loss and 0-1 loss is presented, which proves that the two online Boosting algorithms can maximize the average margin and minimize the margin variance. By estimating the margin mean and variance incrementally, Boosting algorithms can be applied to online learning problems without losing classification accuracy. Experiments on UCI machine learning datasets show that the online Boosting using exponential loss is as accurate as batch AdaBoost, and significantly outperforms the traditional online Boosting, and that the online Boosting using 0-1 loss can minimize classification errors of positive samples and negative samples at the same time, thus applies to imbalance data. Moreover, Boosting using 0-1 loss is more robust on noisy data.
Keywords:AdaBoost  online learning  feature selection  imbalance data
本文献已被 CNKI 等数据库收录!
点击此处可从《自动化学报》浏览原始摘要信息
点击此处可从《自动化学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号