首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Historical Text Normalization with Delayed Rewards
【24h】

Historical Text Normalization with Delayed Rewards

机译:历史文本规范化与延迟奖励

获取原文

摘要

Training neural sequence-to-sequence models with simple token-level log-likelihood is now a standard approach to historical text normalization, albeit often outperformed by phrase-based models. Policy gradient training enables direct optimization for exact matches, and while the small datasets in historical text normalization are prohibitive of from-scratch reinforcement learning, we show that policy gradient fine-tuning leads to significant improvements across the board. Policy gradient training, in particular, leads to more accurate normalizations for long or unseen words.
机译:具有简单令牌级日志可能性的神经序列到序列模型现在是历史文本标准化的标准方法,尽管通常由基于短语的模型表现优于优势。政策梯度训练能够直接优化精确匹配,而历史文本规范中的小型数据集是禁止从头划线学习,我们表明政策梯度微调导致电路板上的显着改进。特别是政策梯度培训,可以为长或看不见的单词导致更准确的训练。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号