【24h】

Using Morphological Knowledge in Open-Vocabulary Neural Language Models

机译:在开放词汇的神经语言模型中使用形态学知识

获取原文

摘要

Languages with productive morphology pose problems for language models that generate words from a fixed vocabulary. Although character-based models allow any possible word type to be generated, they are linguistically naive: they must discover that words exist and are delimited by spaces-basic linguistic facts that are built in to the structure of word-based models. We introduce an open-vocabulary language model that incorporates more sophisticated linguistic knowledge by predicting words using a mixture of three generative processes: (1) by generating words as a sequence of characters, (2) by directly generating full word forms, and (3) by generating words as a sequence of morphemes that are combined using a hand-written morphological analyzer. Experiments on Finnish, Turkish, and Russian show that our model outperforms character sequence models and other strong baselines on intrinsic and extrinsic measures. Furthermore, we show that our model learns to exploit morphological knowledge encoded in the analyzer, and, as a byproduct, it can perform effective unsupervised morphological disambiguation.
机译:具有生产形态的语言会给从固定词汇表生成单词的语言模型带来问题。尽管基于字符的模型允许生成任何可能的单词类型,但是它们在语言上是幼稚的:它们必须发现单词的存在并受到基于单词的模型结构中内置的基于空间的语言事实的界定。我们引入了一种开放词汇的语言模型,该模型通过使用三种生成过程的混合物来预测单词来结合更复杂的语言知识:(1)通过将单词作为字符序列生成;(2)通过直接生成完整的单词形式;以及(3) )生成单词作为语素序列,并使用手写形态分析仪进行组合。在芬兰语,土耳其语和俄语上进行的实验表明,我们的模型优于字符序列模型以及其他有关内在和外在度量的强大基线。此外,我们表明,我们的模型学会了利用分析器中编码的形态学知识,并且作为副产品,它可以执行有效的无监督形态学消歧。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号