【24h】

A Latent Variable Recurrent Neural Network for Discourse Relation Language Models

机译:语篇关系语言模型的隐变量递归神经网络

获取原文

摘要

This paper presents a novel latent variable recurrent neural network architecture for jointly modeling sequences of words and (possibly latent) discourse relations between adjacent sentences. A recurrent neural network generates individual words, thus reaping the benefits of discriminatively-trained vector representations. The discourse relations are represented with a latent variable, which can be predicted or marginalized, depending on the task. The resulting model can therefore employ a training objective that includes not only discourse relation classification, but also word prediction. As a result, it outperforms state-of-the-art alternatives for two tasks: implicit discourse relation classification in the Penn Discourse Treebank, and dialog act classification in the Switchboard corpus. Furthermore, by marginalizing over latent discourse relations at test time, we obtain a discourse informed language model, which improves over a strong LSTM baseline.
机译:本文提出了一种新颖的潜在变量递归神经网络体系结构,用于联合建模单词序列和相邻句子之间的(可能是潜在的)语篇关系。递归神经网络生成单个单词,从而收获了经过区别训练的矢量表示的好处。话语关系用潜在变量表示,根据任务可以预测或边缘化。因此,所得模型可以采用训练目标,该目标不仅包括话语关系分类,还包括单词预测。结果,它优于两项任务的最新选择:宾夕法尼亚州话语树库中的隐式话语关系分类和总机面板语料库中的对话行为分类。此外,通过在测试时边缘化潜在的话语关系,我们获得了话语告知语言模型,该模型在强大的LSTM基线之上得到了改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号