首页> 外文期刊>Future generation computer systems >A lighten CNN-LSTM model for speaker verification on embedded devices
【24h】

A lighten CNN-LSTM model for speaker verification on embedded devices

机译:轻巧的CNN-LSTM模型,用于嵌入式设备上的扬声器验证

获取原文
获取原文并翻译 | 示例
           

摘要

Augmented by deep learning methods, the performance of speaker recognition pipeline has been drastically boosted. For the scenario of smart home, the algorithms of speaker recognition should be user friendly and has high speed, high precision and low resource demand. However, most of the existing algorithms are designed without considering these four performance requirements simultaneously. To fill this gap, this paper proposes a text-independent speaker verification model. Specifically, the lighten network scheme is constructed using one convolution layer, two bilateral Long Short-term Memory (LSTM) layers and one fully connected layer. Utterance segments are mapped to a hypersphere where cosine similarity is used to measure the degree of difference between speakers. Then we analyze the defects of Additive Angular Margin (AAM) loss and propose a 3-stage training method. Softmax pre-training is used for avoiding divergence. After pre-training, AAM loss is adopted to boost training process. In the end, we use triplet loss to further fine-tune the model. Short-term speech utterances are used in training and testing. The experimental results demonstrate that the proposed model reaches 1.17% Equal Error Rate (EER) on a 200 persons benchmark with real-time inference speed on a generic embedded device. (C) 2019 Elsevier B.V. All rights reserved.
机译:通过深度学习方法的增强,说话人识别管道的性能得到了极大提高。对于智能家居场景,说话人识别算法应该是用户友好的,并且具有高速,高精度和低资源需求的特点。但是,大多数现有算法的设计都没有同时考虑这四个性能要求。为了填补这一空白,本文提出了一种与文本无关的说话者验证模型。具体来说,减轻网络方案是使用一个卷积层,两个双边长短期记忆(LSTM)层和一个完全连接的层构造的。话语段被映射到一个超球,其中余弦相似度用于测量说话者之间的差异程度。然后,我们分析了加法角余量(AAM)损失的缺陷,并提出了一种三阶段训练方法。 Softmax预训练用于避免发散。预训练后,采用AAM损失以增强训练过程。最后,我们使用三重态损失来进一步微调模型。短期言语用于训练和测试。实验结果表明,所提出的模型在200人基准上达到了1.17%的平均错误率(EER),并且在通用嵌入式设备上具有实时推理速度。 (C)2019 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号