首页> 外文期刊>Journal of complexity >A proof of convergence for gradient descent in the training of artificial neural networks for constant functions
【24h】

A proof of convergence for gradient descent in the training of artificial neural networks for constant functions

机译:A proof of convergence for gradient descent in the training of artificial neural networks for constant functions

获取原文
获取原文并翻译 | 示例
           

摘要

Gradient descent (GD) optimization algorithms are the standard ingredients that are used to train artificial neural networks (ANNs). However, even in the case of the most basic variant of GD optimization algorithms, the plain vanilla GD method, it remains until today an open problem to prove or disprove the conjecture that GD converges in the training of ANNs. In this article we solve this problem in the special situation where the target function under consideration is a constant function. More specifically, in the case of constant target functions we prove in the training of rectified fully-connected feedforward ANNs with one-hidden layer that the risk function of the GD method does indeed converge to zero. A key contribution of this work is also to explicitly specify a Lyapunov function for the gradient flow system of the ANN parameters. This Lyapunov function is the central tool in our convergence proof. (C) 2022 Elsevier Inc. All rights reserved.

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号