【24h】

Large-scale Multitask Learning for Machine Translation Quality Estimation

机译:大规模多任务学习用于机器翻译质量估计

获取原文

摘要

Multitask learning has been proven a useful technique in a number of Natural Language Processing applications where data is scarce and naturally diverse. Examples include learning from data of different domains and learning from labels provided by multiple annota-tors. Tasks in these scenarios would be the domains or the annotators. When faced with limited data for each task, a framework for the learning of tasks in parallel while using a shared representation is clearly helpful: what is learned for a given task can be transferred to other tasks while the peculiarities of each task are still modelled. Focusing on machine translation quality estimation as application, in this paper we show that multitask learning is also useful in cases where data is abundant. Based on two large-scale datasets, we explore models with multiple annotators and multiple languages and show that state-of-the-art multitask learning algorithms lead to improved results in all settings.
机译:在数据稀缺且自然多样的许多自然语言处理应用中,多任务学习已被证明是一种有用的技术。示例包括从不同域的数据中学习以及从多个注释者提供的标签中学习。这些方案中的任务将是域或注释器。当面对每个任务的有限数据时,使用共享表示形式并行学习任务的框架显然会有所帮助:在仍对每个任务的特性进行建模的同时,可以将为给定任务学到的知识转移到其他任务。重点关注机器翻译质量估计的应用,在本文中,我们表明多任务学习在数据丰富的情况下也很有用。基于两个大型数据集,我们探索了具有多种注释器和多种语言的模型,并显示出最新的多任务学习算法可在所有设置下改善结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号