【24h】

Massively Multilingual Neural Machine Translation

机译:大量多语言神经机翻译

获取原文

摘要

Multilingual neural machine translation (NMT) enables training a single model that supports translation from multiple source languages into multiple target languages. In this paper, we push the limits of multilingual NMT in terms of the number of languages being used. We perform extensive experiments in training massively multilingual NMT models, translating up to 102 languages to and from English within a single model. We explore different setups for training such models and analyze the trade-offs between translation quality and various modeling decisions. We report results on the publicly available TED talks multilingual corpus where we show that massively multilingual many-to-many models are effective in low resource settings, outperforming the previous state-of-the-art while supporting up to 59 languages. Our experiments on a large-scale dataset with 102 languages to and from English and up to one million examples per direction also show promising results, surpassing strong bilingual baselines and encouraging future work on massively multilingual NMT.
机译:多语言神经机翻译(NMT)使训练一个支持从多种源语言的翻译成多种目标语言的模型。在本文中,我们在使用的语言数量方面推动多语言NMT的极限。我们在培训大量多语言NMT型号中进行广泛的实验,在单个模型中翻译高达102种语言和英语。我们探索了培训此类模型的不同设置,并分析翻译质量与各种建模决策之间的权衡。我们向大量语言语料库报告结果,我们显示大量多语言的多语言型号在低资源设置中有效,优于先前的最先进,同时支持最多59种语言。我们在大型数据集上与英语和高达100万个语言的大规模数据集进行的实验,每个方向也表现出了有希望的结果,超越强大的双语基线并鼓励在大量多语言NMT上的未来工作。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号