首页> 外国专利> Stochastic gradient boosting for deep neural networks

Stochastic gradient boosting for deep neural networks

机译:深度神经网络的随机梯度提升

摘要

Aspects described herein may allow for the application of stochastic gradient boosting techniques to the training of deep neural networks by disallowing gradient back propagation from examples that are correctly classified by the neural network model while still keeping correctly classified examples in the gradient averaging. Removing the gradient contribution from correctly classified examples may regularize the deep neural network and prevent the model from overfitting. Further aspects described herein may provide for scheduled boosting during the training of the deep neural network model conditioned on a mini-batch accuracy and/or a number of training iterations. The model training process may start un-boosted, using maximum likelihood objectives or another first loss function. Once a threshold mini-batch accuracy and/or number of iterations are reached, the model training process may begin using boosting by disallowing gradient back propagation from correctly classified examples while continue to average over all mini-batch examples.
机译:本文描述的方面可以允许通过从神经网络模型正确分类的示例禁止梯度回到梯度,以允许随机梯度升压技术应用于深度神经网络的训练,同时仍然保持正确的梯度平均值的正确分类示例。从正确分类的示例中删除渐变贡献可能会规范深度神经网络并防止模型过度拟合。本文描述的其他方面可以在训练期间提供调度提升,在迷你批处理精度和/或多个训练迭代中调节的深神经网络模型。模型培训过程可以使用最大似然目标或另一个第一损耗功能开始未提升。一旦达到了阈值迷你批处理准确度和/或迭代次数,模型训练过程可能会通过禁止从正确的分类示例禁止梯度背部传播而开始使用升级,同时继续平均所有迷你批次示例。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号