首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Multilingual Acoustic Word Embedding Models for Processing Zero-resource Languages
【24h】

Multilingual Acoustic Word Embedding Models for Processing Zero-resource Languages

机译:处理零资源语言的多语言声学词嵌入模型

获取原文

摘要

Acoustic word embeddings are fixed-dimensional representations of variable-length speech segments. In settings where unlabelled speech is the only available resource, such embeddings can be used in "zero-resource" speech search, indexing and discovery systems. Here we propose to train a single supervised embedding model on labelled data from multiple well-resourced languages and then apply it to unseen zero-resource languages. For this transfer learning approach, we consider two multilingual recurrent neural network models: a discriminative classifier trained on the joint vocabularies of all training languages, and a correspondence autoencoder trained to reconstruct word pairs. We test these using a word discrimination task on six target zero-resource languages. When trained on seven well-resourced languages, both models perform similarly and outperform unsupervised models trained on the zero-resource languages. With just a single training language, the second model works better, but performance depends more on the particular training–testing language pair.
机译:声词嵌入是可变长度语音段的固定尺寸表示。在未标记语音是唯一可用资源的设置中,可以在“零资源”语音搜索,索引和发现系统中使用此类嵌入。在这里,我们建议在来自多种资源丰富的语言的标记数据上训练单个监督嵌入模型,然后将其应用于看不见的零资源语言。对于这种转移学习方法,我们考虑两个多语言递归神经网络模型:一个针对所有训练语言的联合词汇进行训练的判别式分类器,以及一个针对重构单词对而进行训练的对应自动编码器。我们使用针对六种目标零资源语言的单词区分任务对它们进行测试。在使用七种资源丰富的语言进行训练时,这两种模型的性能相似,并且优于在零资源语言下进行训练的无监督模型。仅使用一种培训语言,第二种模型就可以更好地工作,但是性能更多地取决于特定的培训-测试语言对。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号