首页> 外文会议>Workshop on Domain Adaptation for NLP >Addressing Zero-Resource Domains Using Document-Level Context in Neural Machine Translation
【24h】

Addressing Zero-Resource Domains Using Document-Level Context in Neural Machine Translation

机译:使用神经计算机翻译中的文档级上下文解决零资源域

获取原文

摘要

Achieving satisfying performance in machine translation on domains for which there is no training data is challenging. Traditional supervised domain adaptation is not suitable for addressing such zero-resource domains because it relies on in-domain parallel data. We show that when in-domain parallel data is not available, access to document-level context enables better capturing of domain generalities compared to only having access to a single sentence. Having access to more information provides a more reliable domain estimation. We present two document-level Transformer models which are capable of using large context sizes and we compare these models against strong Transformer baselines. We obtain improvements for the two zero-resource domains we study. We additionally provide an analysis where we vary the amount of context and look at the case where in-domain data is available.
机译:在没有培训数据的域中实现令人满意的绩效表现,没有培训数据是具有挑战性的。 传统的监督域适应不适合寻址此类零资源域,因为它依赖于域中的并行数据。 我们显示,当域名并行数据不可用时,与只有访问单句话相比,对文档级上下文的访问能够更好地捕获域长。 访问更多信息提供更可靠的域估计。 我们提供了两个能够使用大型上下文尺寸的文件级变压器模型,并将这些模型与强大的变压器基线进行比较。 我们获得了我们学习的两个零资源域的改进。 我们还提供了一个分析,在那里我们改变了上下文的数量,并查看域数据可用的情况。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号