首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Shared-Private Bilingual Word Embeddings for Neural Machine Translation
【24h】

Shared-Private Bilingual Word Embeddings for Neural Machine Translation

机译:用于神经机翻译的共享私人双语词嵌入式

获取原文

摘要

Word embedding is central to neural machine translation (NMT). which has attracted intensive research interest in recent years. In NMT. the source embedding plays the role of the entrance while the target embedding acts as the terminal. These layers occupy most of the model parameters for representation learning. Furthermore, they indirectly interface via a soft-attention mechanism, which makes them comparatively isolated. In this paper, we propose shared-private bilingual word embeddings, which give a closer relationship between the source and target embeddings, and which also reduce the number of model parameters. For similar source and target words, their embeddings tend to share a part of the features and they cooperatively learn these common representation units. Experiments on 5 language pairs belonging to 6 different language families and written in 5 different alphabets demonstrate that the proposed model provides a significant performance boost over the strong baselines with dramatically fewer model parameters.
机译:嵌入词是神经机翻译(NMT)的核心。近年来吸引了密集的研究兴趣。在nmt。源嵌入在目标嵌入作为终端时播放入口的角色。这些层占据了大多数表示学习的模型参数。此外,它们通过软关注机构间接界面,这使得它们相对隔离。在本文中,我们提出了共享的私有双语词嵌入,它在源和目标嵌入之间提供了更紧密的关系,并减少了模型参数的数量。对于类似的来源和目标词语,他们的嵌入倾向于分享特征的一部分,并且它们协同学习这些公共表示单元。 5语言对属于6个不同语言系列的实验,并用5种不同的字母表写入,表明,该模型在强大的基线上提供了显着的性能提升,较为较少的模型参数。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号