【24h】

Interpolated Spectral NGram Language Models

机译:内插光谱ngram语言模型

获取原文

摘要

Spectral models for learning weighted non-deterministic automata have nice theoretical and algorithmic properties. Despite this, it has been challenging to obtain competitive results in language modeling tasks, for two main reasons. First, in order to capture long-range dependencies of the data, the method must use statistics from long substrings, which results in very large matrices that are difficult to decompose. The second is that the loss function behind spectral learning, based on moment matching, differs from the probabilistic metrics used to evaluate language models. In this work we employ a technique for scaling up spectral learning, and use interpolated predictions that are optimized to maximize perplexity. Our experiments in character-based language modeling show that our method matches the performance of state-of-the-art ngram models, while being very fast to train.
机译:用于学习加权非确定性自动机的光谱模型具有很好的理论和算法属性。尽管如此,在语言建模任务中获得竞争结果一直挑战,有两种主要原因。首先,为了捕获数据的远程依赖性,该方法必须使用来自长子字符串的统计信息,这导致难以分解的非常大的矩阵。其次是基于时刻​​匹配的频谱学习背后的损失函数与用于评估语言模型的概率指标不同。在这项工作中,我们采用了一种用于缩放光谱学习的技术,并使用优化以最大化困惑的内插预测。我们在基于性质的语言建模中的实验表明,我们的方法与最先进的ngram模型的性能相匹配,同时非常快速地训练。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号