【24h】

How multilingual is Multilingual BERT?

机译:多语言BERT有多语言能力?

获取原文

摘要

In this paper, we show that Multilingual BERT (M-BERT). released by Devlin et al. (2019) as a single language model pre-trained from monolingual corpora in 104 languages, is surprisingly good at zero-shot cross-lingual model transfer, in which task-specific annotations in one language are used to fine-tune the model for evaluation in another language. To understand why, we present a large number of probing experiments, showing that transfer is possible even to languages in different scripts, that transfer works best between typologically similar languages, that monolingual corpora can train models for code-switching, and that the model can find translation pairs. From these results, we can conclude that M-BERT does create multilingual representations, but that these representations exhibit systematic deficiencies affecting certain language pairs.
机译:在本文中,我们展示了多语言BERT(M-BERT)。由Devlin等人发布。 (2019)是从104种语言的单语语料库预训练的一种语言模型,令人惊讶地擅长于零镜头跨语言模型传递,其中使用一种语言的特定于任务的注释来微调模型以进行评估用另一种语言。为了理解原因,我们提出了大量的探测实验,表明即使在不同脚本中的语言也可以进行转移,在类型相似的语言之间转移效果最好,单语语料库可以训练模型进行代码转换,并且模型可以查找翻译对。从这些结果,我们可以得出结论,M-BERT确实创建了多语言表示形式,但是这些表示形式表现出影响某些语言对的系统缺陷。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号