【24h】

Attention Word Embedding

机译:注意词嵌入

获取原文

摘要

Word embedding models learn semantically rich vector representations of words and are widely used to initialize natural processing language (NLP) models. The popular continuous bag-of-words (CBOW) model of word2vec learns a vector embedding by masking a given word in a sentence and then using the other words as a context to predict it. A limitation of CBOW is that it equally weights the context words when making a prediction, which is inefficient, since some words have higher predictive value than others. We tackle this inefficiency by introducing the Attention Word Embedding (AWE) model, which integrates the attention mechanism into the CBOW model. We also propose AWE-S, which incorporates subword information. We demonstrate that AWE and AWE-S outperform the state-of-the-art word embedding models both on a variety of word similarity datasets and when used for initialization of NLP models.
机译:Word嵌入式模型学习语义上丰富的单词矢量表示,并广泛用于初始化自然处理语言(NLP)模型。 Word2Vec的流行连续单词(CBOW)模型通过屏蔽句子中的给定单词来了解嵌入的向量嵌入,然后使用其他单词作为预测它的上下文。 Cow的限制是它在制造预测时,它同样重量上下文词,这是效率低下的,因为一些词具有比其他单词更高的预测值。 我们通过引入注意力嵌入(敬畏)模型来解决这种效率,这将注意力机制集成到CBY模型中。 我们还提出了敬畏的,它包含子字信息。 我们展示了敬畏和敬畏,效果均在各种单词相似性数据集上且用于初始化NLP模型时的最先进的单词嵌入模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号