【24h】

A Diversity-Promoting Objective Function for Neural Conversation Models

机译:神经对话模型的促进多样性的目标函数

获取原文

摘要

Sequence-to-sequence neural network models for generation of conversational responses tend to generate safe, commonplace responses (e.g., I don't know) regardless of the input. We suggest that the traditional objective function, i.e., the likelihood of output (response) given input (message) is unsuited to response generation tasks. Instead we propose using Maximum Mutual Information (MMI) as the objective function in neural models. Experimental results demonstrate that the proposed MMI models produce more diverse, interesting, and appropriate responses, yielding substantive gains in Bleu scores on two conversational datasets and in human evaluations.
机译:无论输入如何,用于生成会话响应的序列到序列神经网络模型都倾向于生成安全,普通的响应(例如,我不知道)。我们建议传统的目标函数,即给定输入(消息)的输出(响应)的可能性不适合响应生成任务。相反,我们建议在神经模型中使用最大互信息(MMI)作为目标函数。实验结果表明,所提出的MMI模型产生了更加多样化,有趣且适当的响应,从而在两个会话数据集和人类评估中的Bleu分数上取得了实质性的收益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号