...
首页> 外文期刊>Entropy >Simple Stopping Criteria for Information Theoretic Feature Selection
【24h】

Simple Stopping Criteria for Information Theoretic Feature Selection

机译:信息理论特征选择的简单停止准则

获取原文
           

摘要

Feature selection aims to select the smallest feature subset that yields the minimum generalization error. In the rich literature in feature selection, information theory-based approaches seek a subset of features such that the mutual information between the selected features and the class labels is maximized. Despite the simplicity of this objective, there still remain several open problems in optimization. These include, for example, the automatic determination of the optimal subset size (i.e., the number of features) or a stopping criterion if the greedy searching strategy is adopted. In this paper, we suggest two stopping criteria by just monitoring the conditional mutual information (CMI) among groups of variables. Using the recently developed multivariate matrix-based Rényi’s α -entropy functional, which can be directly estimated from data samples, we showed that the CMI among groups of variables can be easily computed without any decomposition or approximation, hence making our criteria easy to implement and seamlessly integrated into any existing information theoretic feature selection methods with a greedy search strategy.
机译:特征选择旨在选择产生最小泛化误差的最小特征子集。在有关特征选择的丰富文献中,基于信息论的方法寻求特征的子集,以使所选特征和类别标签之间的相互信息最大化。尽管此目标很简单,但优化中仍然存在一些未解决的问题。这些包括例如自动确定最佳子集大小(即,特征数量)或如果采用贪婪搜索策略则停止准则。在本文中,我们仅通过监视变量组之间的条件互信息(CMI)来建议两个停止标准。使用最近开发的基于多元矩阵的Rényiα熵函数(可以直接从数据样本中进行估计),我们表明变量组之间的CMI可以轻松计算而无需任何分解或近似,因此使我们的标准易于实现和通过贪婪搜索策略无缝地集成到任何现有的信息理论特征选择方法中。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号