首页> 外文会议>International Conference on Computational Linguistics >Language Model Transformers as Evaluators for Open-domain Dialogues
【24h】

Language Model Transformers as Evaluators for Open-domain Dialogues

机译:语言模型变压器作为开放式对话的评估员

获取原文

摘要

Computer-based systems for communication with humans are a cornerstone of AI research since the 1950s. So far, the most effective way to assess the quality of the dialogues produced by these systems is to use resource-intensive manual labor instead of automated means. In this work, we investigate whether language models (LM) based on transformer neural networks can indicate the quality of a conversation. In a general sense, language models are methods that learn to predict one or more words based on an already given context. Due to their unsupervised nature, they are candidates for efficient, automatic indication of dialogue quality. We demonstrate that human evaluators have a positive correlation between the output of the language models and scores. We also provide some insights into their behavior and inner-working in a conversational context.
机译:自20世纪50年代以来,基于计算机的通信系统是AI研究的基石。 到目前为止,评估这些系统产生的对话质量的最有效的方法是使用资源密集的手工劳动而不是自动化手段。 在这项工作中,我们调查了基于变压器神经网络的语言模型(LM)是否可以指示对话的质量。 在一般意义上,语言模型是学习基于已经给定的上下文预测一个或多个单词的方法。 由于他们无人监督的性质,他们是有效,自动指示对话质量的候选人。 我们证明人类评估人员在语言模型和分数的产出之间具有正相关。 我们还向其行为提供了一些见解和在会话环境中的内心工作。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号