【24h】

Bayesian Learning for Neural Dependency Parsing

机译:贝叶斯学习神经依赖解析

获取原文

摘要

While neural dependency parsers provide state-of-the-art accuracy for several languages, they still rely on large amounts of costly labeled training data. We demonstrate that in the small data regime, where uncertainty around parameter estimation and model prediction matters the most, Bayesian neural modeling is very effective. In order to overcome the computational and statistical costs of the approximate inference step in this framework, we utilize an efficient sampling procedure via stochastic gradient Langevin dynamics to generate samples from the approximated posterior. Moreover, we show that our Bayesian neural parser can be further improved when integrated into a multi-task parsing and POS tagging framework, designed to minimize task interference via an adversarial procedure. When trained and tested on 6 languages with less than 5k training instances, our parser consistently outperforms the strong BiLSTM baseline (Kiper-wasser and Goldberg, 2016). Compared with the BiAFFINE parser (Dozat et al., 2017) our model achieves an improvement of up to 3% for Vietnamese and Irish, while our multi-task model achieves an improvement of up to 9% across five languages: Farsi, Russian, Turkish. Vietnamese, and Irish.
机译:虽然神经依赖解析器为多种语言提供最先进的准确性,但它们仍然依靠大量的昂贵标记的训练数据。我们证明,在小数据制度中,在参数估计和模型预测周围的不确定性最重要的情况下,贝叶斯神经建模非常有效。为了克服本框架中的近似推理步骤的计算和统计成本,我们通过随机梯度Langevin动力学利用有效的采样过程来从近似后的后部产生样品。此外,我们表明,当集成到多任务解析和POS标记框架中时,我们的贝叶斯神经解析器可以进一步提高,旨在通过对抗过程最小化任务干扰。当培训和测试6种语言的培训实例,我们的解析器始终如一地优于强大的Bilstm基线(Kiper-Wasser和Goldberg,2016)。与Biaffine Parser(Dozat等,2017)相比,我们的模式实现了越南和爱尔兰最高可达3%的增长,而我们的多任务模型可实现五种语言的9%:Farsi,俄罗斯,土耳其。越南人,爱尔兰人。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号