首页> 外国专利> Stochastic Gradient Boosting For Deep Neural Networks

Stochastic Gradient Boosting For Deep Neural Networks

机译:深度神经网络的随机梯度提升

摘要

Aspects described herein may allow for the application of stochastic gradient boosting techniques to the training of deep neural networks by disallowing gradient back propagation from examples that are correctly classified by the neural network model while still keeping correctly classified examples in the gradient averaging. Removing the gradient contribution from correctly classified examples may regularize the deep neural network and prevent the model from overfitting. Further aspects described herein may provide for scheduled boosting during the training of the deep neural network model conditioned on a mini-batch accuracy and/or a number of training iterations. The model training process may start un-boosted, using maximum likelihood objectives or another first loss function. Once a threshold mini-batch accuracy and/or number of iterations are reached, the model training process may begin using boosting by disallowing gradient back propagation from correctly classified examples while continue to average over all mini-batch examples.
机译:本文所述的方面可以通过允许来自由神经网络模型正确分类的示例的梯度反向传播,而仍保留梯度平均中正确分类的示例,从而允许将随机梯度增强技术应用于深度神经网络的训练。从正确分类的示例中删除梯度贡献可能会规范深度神经网络,并防止模型过度拟合。本文所述的其他方面可以在以小批量精度和/或多次训练迭代为条件的深度神经网络模型的训练期间提供计划的增强。使用最大似然目标或另一个第一损失函数,模型训练过程可以不加任何假设地开始。一旦达到阈值小批量精度和/或迭代次数,模型训练过程就可以通过禁止来自正确分类的示例的梯度反向传播而开始使用增强,同时继续对所有小批量示例求平均。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号