首页> 外文会议>Workshop on Commonsense inference in natural language processing >When Choosing Plausible Alternatives, Clever Hans can be Clever
【24h】

When Choosing Plausible Alternatives, Clever Hans can be Clever

机译:当选择可行的替代品,聪明汉斯可以巧妙

获取原文

摘要

Pretrained language models, such as BERT and RoBERTa, have shown large improvements in the commonsense reasoning benchmark COPA. However, recent work found that many improvements in benchmarks of natural language understanding are not due to models learning the task, but due to their increasing ability to exploit superficial cues, such as tokens that occur more often in the correct answer than the wrong one. Are BERT's and RoBERTa's good performance on COPA also caused by this? We find superficial cues in COPA, as well as evidence that BERT exploits these cues. To remedy this problem, we introduce Balanced COPA, an extension of COPA that does not suffer from easy-to-exploit single token cues. We analyze BERT's and RoBERTa's performance on original and Balanced COPA, finding that BERT relies on superficial cues when they are present, but still achieves comparable performance once they arc made ineffective, suggesting that BERT learns the task to a certain degree when forced to. In contrast, RoBERTa does not appear to rely on superficial cues.
机译:诸如BERT和RoBERTa之类的经过预训练的语言模型在常识推理基准COPA中显示出了很大的改进。但是,最近的工作发现,自然语言理解基准的许多改进不是由于学习任务的模型,而是由于它们利用表象线索的能力不断提高,例如在正确答案中出现的令牌比在错误答案中出现的频率高。 BERT和RoBERTa在COPA上的良好表现是否也是由这引起的?我们在COPA中找到了肤浅的线索,也有证据表明BERT利用了这些线索。为了解决这个问题,我们引入了Balanced COPA,它是COPA的扩展,它没有易于利用的单个令牌提示。我们分析了BERT和RoBERTa在原始COPA和Balanced COPA上的表现,发现BERT在出现时依赖于肤浅的提示,但是一旦它们变得无效,仍然可以达到可比的性能,这表明BERT在被迫执行时可以在一定程度上学习任务。相比之下,RoBERTa似乎并不依赖肤浅的线索。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号