【24h】

Learning to Rank for Plausible Plausibility

机译:学习为合理性排名

获取原文

摘要

Researchers illustrate improvements in contextual encoding strategies via resultant performance on a battery of shared Natural Language Understanding (NLU) tasks. Many of these tasks are of a categorical prediction variety: given a conditioning context (e.g., an NLI premise), provide a label based on an associated prompt (e.g., an NLI hypothesis). The categorical nature of these tasks has led to common use of a cross entropy log-loss objective during training. We suggest this loss is intuitively wrong when applied to plausibility tasks, where the prompt by design is neither categorically entailed nor contradictory given the context. Log-loss naturally drives models to assign scores near 0.0 or 1.0, in contrast to our proposed use of a margin-based loss. Following a discussion of our intuition, we describe a confirmation study based on an extreme, synthetically curated task derived from MultiNLI. We find that a margin-based loss leads to a more plausible model of plausibility. Finally, we illustrate improvements on the Choice Of Plausible Alternative (COPA) task through this change in loss.
机译:研究人员通过共享自然语言理解(NLU)任务的综合性能说明了上下文编码策略的改进。这些任务中的许多任务都是分类预测的:给定条件上下文(例如,NLI前提),根据相关提示(例如,NLI假设)提供标签。这些任务的分类性质导致在训练过程中普遍使用交叉熵对数损失物镜。我们建议,将这种损失应用于合理性任务时,从直觉上来说是错误的,在这种情况下,设计提示既不是绝对必要的,也不是给定上下文的矛盾。对数损失自然会驱使模型将分数分配到0.0或1.0附近,这与我们建议的基于保证金的损失的使用形成对比。在讨论了我们的直觉之后,我们描述了基于从MultiNLI派生出来的一项极端的,精心策划的任务进行的确认研究。我们发现基于保证金的损失导致了更合理的合理性模型。最后,我们通过这种损失变化说明了对“合理选择方案”(COPA)任务的改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号