【24h】

Recurrent Memory Networks for Language Modeling

机译:循环内存网络用于语言建模

获取原文

摘要

Recurrent Neural Networks (RNNs) have obtained excellent result in many natural language processing (NLP) tasks. However, understanding and interpreting the source of this success remains a challenge. In this paper, we propose Recurrent Memory Network (RMN), a novel RNN architecture, that not only amplifies the power of RNN but also facilitates our understanding of its internal functioning and allows us to discover underlying patterns in data. We demonstrate the power of RMN on language modeling and sentence completion tasks. On language modeling, RMN outperforms Long Short-Term Memory (LSTM) network on three large German, Italian, and English dataset. Additionally we perform in-depth analysis of various linguistic dimensions that RMN captures. On Sentence Completion Challenge, for which it is essential to capture sentence coherence, our RMN obtains 69.2% accuracy, surpassing the previous state of the art by a large margin.
机译:递归神经网络(RNN)在许多自然语言处理(NLP)任务中均获得了出色的成绩。然而,理解和解释这种成功的根源仍然是一个挑战。在本文中,我们提出了一种循环神经网络(RMN),这是一种新颖的RNN体系结构,它不仅可以放大RNN的功能,还可以帮助我们了解其内部功能,并允许我们发现数据中的潜在模式。我们展示了RMN在语言建模和句子完成任务上的强大功能。在语言建模方面,RMN在三个大型德语,意大利语和英语数据集上的表现优于长期短期记忆(LSTM)网络。此外,我们对RMN捕获的各种语言维度进行深入分析。在完成句子连贯性至关重要的“句子完成挑战”上,我们的RMN可获得69.2%的准确率,大大超过了现有技术水平。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号