首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Towards Robust and Privacy-preserving Text Representations
【24h】

Towards Robust and Privacy-preserving Text Representations

机译:朝着强大而隐私保留的文本表示

获取原文

摘要

Written text often provides sufficient clues to identify the author, their gender, age, and other important attributes. Consequently, the authorship of training and evaluation corpora can have unforeseen impacts, including differing model performance for different user groups, as well as privacy implications. In this paper, we propose an approach to explicitly obscure important author characteristics at training time, such that representations learned are invariant to these attributes. Evaluating on two tasks, we show that this leads to increased privacy in the learned representations, as well as more robust models to varying evaluation conditions, including out-of-domain corpora.
机译:书面文本经常提供足够的线索来识别作者,其性别,年龄和其他重要属性。因此,培训和评估的作者可以具有无法预料的影响,包括不同用户组的不同模式性能,以及隐私含义。在本文中,我们提出了一种在培训时间明确地模糊重要作者特征的方法,使得学到的表示是不变的这些属性。评估两项任务,我们表明这导致了解到所学习的陈述中的隐私,以及更强大的模型,以不同的评估条件,包括域名基础。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号