【24h】

Weighting Finite-State Transductions With Neural Context

机译:带有神经环境的有限状态转换加权

获取原文

摘要

How should one apply deep learning to tasks such as morphological reinflection, which stochastically edit one string to get another? A recent approach to such sequence-to-sequence tasks is to compress the input string into a vector that is then used to generate the output string, using recurrent neural networks. In contrast, we propose to keep the traditional architecture, which uses a finite-state transducer to score all possible output strings, but to augment the scoring function with the help of recurrent networks. A stack of bidirectional LSTMs reads the input string from left-to-right and right-to-left, in order to summarize the input context in which a transducer arc is applied. We combine these learned features with the transducer to define a probability distribution over aligned output strings, in the form of a weighted finite-state automaton. This reduces hand-engineering of features, allows learned features to examine unbounded context in the input string, and still permits exact inference through dynamic programming. We illustrate our method on the tasks of morphological reinflection and lemmatization.
机译:人们应该如何将深度学习应用于诸如形态学改造之类的任务,这些任务会随机编辑一个字符串以获得另一个字符串?用于此类序列到序列任务的最新方法是使用递归神经网络将输入字符串压缩为向量,然后将其用于生成输出字符串。相反,我们建议保留传统架构,该架构使用有限状态换能器对所有可能的输出字符串进行评分,但在递归网络的帮助下增加评分功能。双向LSTM堆栈从左到右和从右到左读取输入字符串,以便总结应用换能器电弧的输入上下文。我们将这些学习到的特征与换能器相结合,以加权有限状态自动机的形式定义对齐的输出字符串上的概率分布。这减少了对特征的手工操作,允许学习的特征检查输入字符串中的无限制上下文,并且仍然允许通过动态编程进行精确推断。我们举例说明了我们在形态学改形和词形化任务上的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号