【24h】

Neural Self-Training through Spaced Repetition

机译:通过间隔重复的神经训练

获取原文

摘要

Self-training is a semi-supervised learning approach for utilizing unlabeled data to create better learners. The efficacy of self-training algorithms depends on their data sampling techniques. The majority of current sampling techniques are based on predetermined policies which may not effectively explore the data space or improve model generalizability. In this work, we tackle the above challenges by introducing a new data sampling technique based on spaced repetition that dynamically samples informative and diverse unlabeled instances with respect to individual learner and instance characteristics. The proposed model is specifically effective in the context of neural models which can suffer from overfitting and high-variance gradients when trained with small amount of labeled data. Our model outperforms current semi-supervised learning approaches developed for neural networks on publicly-available datasels.
机译:自我培训是一种半监督的学习方法,用于利用未标记的数据来创造更好的学习者。自我训练算法的功效取决于其数据采样技术。大多数当前采样技术基于预定的策略,这可能不会有效地探索数据空间或改善型号的概括性。在这项工作中,我们通过基于间隔重复引入新的数据采样技术来解决上述挑战,以动态地示出了各个学习者和实例特征的信息和不同的未标记实例。该建议的模型在神经模型的背景下特别有效,当培训时,在具有少量标记数据的训练时可能受到过度拟合和高方差梯度。我们的模型优于用于公开数据的神经网络开发的目前的半监督学习方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号