【24h】

Multi-domain Neural Network Language Generation for Spoken Dialogue Systems

机译:语音对话系统的多域神经网络语言生成

获取原文

摘要

Moving from limited-domain natural language generation (NLG) to open domain is difficult because the number of semantic input combinations grows exponentially with the number of domains. Therefore, it is important to leverage existing resources and exploit similarities between domains to facilitate domain adaptation. In this paper, we propose a procedure to train multi-domain, Recurrent Neural Network-based (RNN) language generators via multiple adaptation steps. In this procedure, a model is first trained on counterfeited data synthesised from an out-of-domain dataset, and then fine tuned on a small set of in-domain utterances with a discriminative objective function. Corpus-based evaluation results show that the proposed procedure can achieve competitive performance in terms of BLEU score and slot error rate while significantly reducing the data needed to train generators in new, unseen domains. In subjective testing, human judges confirm that the procedure greatly improves generator performance when only a small amount of data is available in the domain.
机译:从有限域自然语言生成(NLG)到开放域是困难的,因为语义输入组合的数量随域的数量呈指数增长。因此,利用现有资源并利用域之间的相似性以促进域适应很重要。在本文中,我们提出了一种通过多个适应步骤来训练基于多域递归神经网络(RNN)的语言生成器的过程。在此过程中,首先对从域外数据集合成的伪造数据进行模型训练,然后对具有判别目标函数的一小部分域内话语进行微调。基于语料库的评估结果表明,所提出的程序可以在BLEU分数和时隙错误率方面取得竞争优势,同时显着减少在新的,看不见的领域中训练生成器所需的数据。在主观测试中,人类法官确认,当域中只有少量数据可用时,该程序将大大提高生成器的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号