...
【24h】

Double committee adaboost

机译:双重委员会

获取原文
           

摘要

In this paper we make an extensive study of different combinations of ensemble techniques for improving the performance of adaboost considering the following strategies: reducing the correlation problem among the features, reducing the effect of the outliers in adaboost training, and proposing an efficient way for selecting/weighing the weak learners. First, we show that random subspace works well coupled with several adaboost techniques. Second, we show that an ensemble based on training perturbation using editing methods (to reduce the importance of the outliers) further improves performance. We examine the robustness of the new approach by applying it to a number of benchmark datasets representing a range of different problems. We find that compared with other state-of-the-art classifiers our proposed method performs consistently well across all the tested datasets. One useful finding is that this approach obtains a performance similar to support vector machine (SVM), using the well-known LibSVM implementation, even when both kernel selection and various parameters of SVM are carefully tuned for each dataset. The main drawback of the proposed approach is the computation time, which is high as a result of combining the different ensemble techniques. We have also tested the fusion between our selected committee of adaboost with SVM (again using the widely tested LibSVM tool) where the parameters of SVM are tuned for each dataset. We find that the fusion between SVM and a committee of adaboost (i.e., a heterogeneous ensemble) statistically outperforms the most used SVM tool with parameters tuned for each dataset. The MATLAB code of our best approach is available at bias.csr.unibo.itanni/ADA.rar .
机译:在本文中,我们通过考虑以下策略对用于提高adaboost性能的不同合奏技术组合进行了广泛研究:减少特征之间的相关性问题,减少adaboost训练中离群值的影响,并提出一种有效的选择方法/权衡弱小的学习者。首先,我们证明随机子空间与几种adaboost技术配合使用效果很好。其次,我们证明了使用编辑方法(以减少异常值的重要性)为基础的基于训练摄动的合奏可以进一步提高性能。我们通过将新方法应用于代表一系列不同问题的许多基准数据集来检验新方法的鲁棒性。我们发现,与其他最新分类器相比,我们提出的方法在所有测试数据集上的性能始终如一。一个有用的发现是,即使为每个数据集仔细调整了内核选择和SVM的各种参数,该方法也可以使用众所周知的LibSVM实现获得与支持向量机(SVM)相似的性能。所提出的方法的主要缺点是计算时间,由于结合了不同的集成技术,因此计算时间很高。我们还测试了选定的adaboost委员会与SVM之间的融合(再次使用经过广泛测试的LibSVM工具),其中针对每个数据集调整了SVM的参数。我们发现,在支持向量机和adaboost委员会(即异构集合)之间的融合在统计上胜过了最常用的支持向量机工具,其参数针对每个数据集进行了调整。我们最好的方法的MATLAB代码可从bias.csr.unibo.itanni/ADA.rar获得。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号