首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Attention Is (not) All You Need for Commonsense Reasoning
【24h】

Attention Is (not) All You Need for Commonsense Reasoning

机译:注意不是常识推理所需的全部

获取原文

摘要

The recently introduced BERT model exhibits strong performance on several language understanding benchmarks. In this paper, we describe a simple re-implementation of BERT for commonsense reasoning. We show that the attentions produced by BERT can be directly utilized for tasks such as the Pronoun Disambiguation Problem and Winograd Schema Challenge. Our proposed attention-guided commonsense reasoning method is conceptually simple yet empirically powerful. Experimental analysis on multiple datasets demonstrates that our proposed system performs remarkably well on all cases while outperforming the previously reported state of the art by a margin. While results suggest that BERT seems to implicitly learn to establish complex relationships between entities, solving commonsense reasoning tasks might require more than unsupervised models learned from huge text corpora.
机译:最近推出的BERT模型在几种语言理解基准上均表现出出色的性能。在本文中,我们描述了用于常识推理的BERT的简单重新实现。我们证明了BERT产生的注意力可以直接用于代词消歧问题和Winograd模式挑战之类的任务。我们提出的注意导向的常识推理方法在概念上很简单,但从经验上讲却很有效。对多个数据集的实验分析表明,我们提出的系统在所有情况下均具有出色的性能,同时在一定程度上优于先前报告的最新技术水平。虽然结果表明BERT似乎暗中学习了建立实体之间的复杂关系,但解决常识推理任务可能比从大型文本语料库中学习的无监督模型需要更多的东西。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号