首页> 外文学位 >A Comparison of the Quality of Rule Induction from Inconsistent Data Sets and Incomplete Data Sets.
【24h】

A Comparison of the Quality of Rule Induction from Inconsistent Data Sets and Incomplete Data Sets.

机译:来自不一致数据集和不完整数据集的规则归纳质量的比较。

获取原文
获取原文并翻译 | 示例

摘要

In data mining, decision rules induced from known examples are used to classify unseen cases. There are various rule induction algorithms, such as LEM1 (Learning from Examples Module version 1), LEM2 (Learning from Examples Module version 2) and MLEM2 (Modified Learning from Examples Module version 2). In the real world, many data sets are imperfect, either inconsistent or incomplete. The idea of lower and upper approximations, or more generally, the probabilistic approximation, provides an effective way to induce rules from inconsistent data sets and incomplete data sets. But the accuracies of rule sets induced from imperfect data sets are expected to be lower. The objective of this project is to investigate which kind of imperfect data sets (inconsistent or incomplete) is worse in terms of the quality of rule induction. In this project, experiments were conducted on eight inconsistent data sets and eight incomplete data sets with lost values. We implemented the MLEM2 algorithm to induce certain and possible rules from inconsistent data sets, and implemented the local probabilistic version of MLEM2 algorithm to induce certain and possible rules from incomplete data sets. A program called Rule Checker was also developed to classify unseen cases with induced rules and measure the classification error rate. Ten-fold cross validation was carried out and the average error rate was used as the criterion for comparison. The Mann-Whitney nonparametric tests were performed to compare, separately for certain and possible rules, incompleteness with inconsistency. The results show that there is no significant difference between inconsistent and incomplete data sets in terms of the quality of rule induction.
机译:在数据挖掘中,将从已知示例中得出的决策规则用于对未发现的案例进行分类。有各种规则归纳算法,例如LEM1(从示例模块版本1学习),LEM2(从示例模块版本2学习)和MLEM2(从示例模块版本2学习)。在现实世界中,许多数据集是不完美的,不一致或不完整。上下近似,或更一般地说,概率近似的思想提供了一种有效的方法,可以从不一致的数据集和不完整的数据集推导规则。但是,由不完善的数据集引起的规则集的准确性预计会更低。该项目的目的是研究哪种不完善的数据集(不一致或不完整)在规则归纳的质量方面更差。在该项目中,对八个不一致的数据集和八个具有丢失值的不完整数据集进行了实验。我们实现了MLEM2算法以从不一致的数据集中导出某些和可能的规则,并实现了MLEM2算法的本地概率版本,以从不完整的数据集中导出某些和可能的规则。还开发了一个名为Rule Checker的程序,用于使用诱导规则对看不见的案例进行分类,并测量分类错误率。进行十次交叉验证,并将平均错误率用作比较标准。进行了Mann-Whitney非参数检验,以分别比较某些规则和可能的规则的不完整性和不一致性。结果表明,就规则归纳的质量而言,不一致和不完整的数据集之间没有显着差异。

著录项

  • 作者

    Su, Xiaomeng.;

  • 作者单位

    University of Kansas.;

  • 授予单位 University of Kansas.;
  • 学科 Computer science.
  • 学位 M.S.
  • 年度 2015
  • 页码 53 p.
  • 总页数 53
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号