【24h】

Ultradense Word Embeddings by Orthogonal Transformation

机译:正交变换的超密集词嵌入

获取原文

摘要

Embeddings are generic representations that are useful for many NLP tasks. In this paper, we introduce Densifier, a method that learns an orthogonal transformation of the embedding space that focuses the information relevant for a task in an ultradense subspace of a dimensionality that is smaller by a factor of 100 than the original space. We show that ultradense embeddings generated by Densifier reach state of the art on a lexicon creation task in which words are annotated with three types of lexical information - sentiment, con-creteness and frequency. On the SemEval2015 10B sentiment analysis task we show that no information is lost when the ultradense sub-space is used, but training is an order of magnitude more efficient due to the compactness of the ultradense space.
机译:嵌入是可用于许多NLP任务的通用表示形式。在本文中,我们介绍了Densifier,它是一种学习嵌入空间的正交变换的方法,该方法将与任务相关的信息集中在维数比原始空间小100倍的超密集子空间中。我们表明,由Densifier生成的超密集嵌入在词汇创建任务上达到了最新水平,在该任务中用三种类型的词汇信息(情感,具体程度和频率)对单词进行注释。在SemEval2015 10B情感分析任务上,我们显示了使用超密集子空间时不会丢失任何信息,但是由于超密集空间的紧凑性,训练的效率提高了一个数量级。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号