首页> 外文会议> >An Embarrassingly Simple Approach for Transfer Learning from Pretrained Language Models
【24h】

An Embarrassingly Simple Approach for Transfer Learning from Pretrained Language Models

机译:从预训练的语言模型进行迁移学习的一种非常尴尬的简单方法

获取原文

摘要

A growing number of state-of-the-art transfer learning methods employ language models pretrained on large generic corpora. In this paper we present a conceptually simple and effective transfer learning approach that addresses the problem of catastrophic forgetting. Specifically, we combine the task-specific optimization function with an auxiliary language model objective, which is adjusted during the training process. This preserves language regularities captured by language models, while enabling sufficient adaptation for solving the target task. Our method does not require pre-training or finetuning separate components of the network and we train our models end-to-end in a single step. We present results on a variety of challenging affective and text classification tasks, surpassing well established transfer learning methods with greater level of complexity.
机译:越来越多的最新转移学习方法采用在大型通用语料库上预先训练的语言模型。在本文中,我们提出了一种概念上简单有效的转移学习方法,用于解决灾难性遗忘问题。具体来说,我们将特定于任务的优化功能与辅助语言模型目标结合在一起,该目标可以在培训过程中进行调整。这样可以保留语言模型捕获的语言规则,同时可以进行充分的调整以解决目标任务。我们的方法不需要预先训练或微调网络的各个组成部分,而我们只需一步就可以对模型进行端到端训练。我们介绍了各种具有挑战性的情感和文本分类任务的结果,这些结果超过了成熟的,具有更高复杂度的转移学习方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号