首页> 外文会议>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP Models
【24h】

Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP Models

机译:注意中毒词嵌入式:探索NLP模型中嵌入层的脆弱性

获取原文

摘要

Recent studies have revealed a security threat to natural language processing (NLP) models, called the Backdoor Attack. Victim models can maintain competitive performance on clean samples while behaving abnormally on samples with a specific trigger word inserted. Previous backdoor attacking methods usually assume that attackers have a certain degree of data knowledge, either the dataset which users would use or proxy datasets for a similar task, for implementing the data poisoning procedure. However, in this paper, we find that it is possible to hack the model in a data-free way by modifying one single word embedding vector, with almost no accuracy sacrificed on clean samples. Experimental results on sentiment analysis and sentence-pair classification tasks show that our method is more efficient and stealthier. We hope this work can raise the awareness of such a critical security risk hidden in the embedding layers of NLP models.
机译:最近的研究揭示了对自然语言处理(NLP)模型的安全威胁,称为后门攻击。 受害者模型可以在清洁样本上保持竞争性能,同时在具有插入特定触发单词的样本上表现异常。 以前的后门攻击方法通常假设攻击者具有一定程度的数据知识,该数据集可以使用或代理数据集进行类似的任务,用于实现数据中毒过程。 然而,在本文中,我们发现通过修改一个单词嵌入向量,可以以无数据的方式破解模型,几乎没有在清洁样品上牺牲的准确度。 关于情感分析和句子分类任务的实验结果表明,我们的方法更有效,悄悄。 我们希望这项工作能够提高隐藏在NLP模型的嵌入层中这种关键安全风险的意识。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号