首页> 外文会议>Conference on the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >Ontologically Grounded Multi-sense Representation Learning for Semantic Vector Space Models
【24h】

Ontologically Grounded Multi-sense Representation Learning for Semantic Vector Space Models

机译:用于语义矢量空间模型的本体地面接地的多感官表示学习

获取原文

摘要

Words are polysemous. However, most approaches to representation learning for lexical semantics assign a single vector to every surface word type. Meanwhile, lexical ontologies such as WordNet provide a source of complementary knowledge to distributional information, including a word sense inventory. In this paper we propose two novel and general approaches for generating sense-specific word embeddings that are grounded in an ontology. The first applies graph smoothing as a postprocessing step to tease the vectors of different senses apart, and is applicable to any vector space model. The second adapts predictive maximum likelihood models that learn word embeddings with latent variables representing senses grounded in an specified ontology. Empirical results on lexical semantic tasks show that our approaches effectively captures information from both the ontology and distributional statistics. Moreover, in most cases our sense-specific models outperform other models we compare against.
机译:言语是多园。然而,大多数对词汇语义的表示学习的方法为每个表面字类型分配了一个向量。同时,Wordnet等词法本体提供了与分布信息的互补知识来源,包括单词感测库存。在本文中,我们提出了两种新颖的和一般方法,用于生成在本体中接地的感觉特定的单词嵌入。第一种应用图表平滑作为后处理步骤,以使不同感官的矢量分开,并且适用于任何向量空间模型。第二种适应预测的最大似然模型,了解具有表示在指定本体的潜在变量的词嵌入词嵌入式。词汇语义任务的经验结果表明,我们的方法有效地捕获了本体和分布统计的信息。此外,在大多数情况下,我们的感觉特定模型优于我们比较的其他模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号