首页> 外文会议>International Conference on Machine Learning >Dynamics of Deep Neural Networks and Neural Tangent Hierarchy
【24h】

Dynamics of Deep Neural Networks and Neural Tangent Hierarchy

机译:深神经网络和神经切线等级的动态

获取原文

摘要

The evolution of a deep neural network trained by the gradient descent in the overparametrization regime can be described by its neural tangent kernel (NTK) (Jacot et al., 2018; Du et al., 2018b;a; Arora et al., 2019b). It was observed (Arora et al., 2019a) that there is a performance gap between the kernel regression using the limiting NTK and the deep neural networks. We study the dynamic of neural networks of finite width and derive an infinite hierarchy of differential equations, the neural tangent hierarchy (NTH). We prove that the NTH hierarchy truncated at the level p ≥ 2 approximates the dynamic of the NTK up to arbitrary precision under certain conditions on the neural network width and the data set dimension. The assumptions needed for these approximations become weaker as p increases. Finally, NTH can be viewed as higher order extensions of NTK. In particular, the NTH truncated at p = 2 recovers the NTK dynamics.
机译:通过梯度下降在超分度化方案中训练的深神经网络的演变可以由其神经切线内核(NTK)(Jacot等,2018; Du等人,2018b; A; Arora等,2019b )。 观察到(Arora等,2019a),使用限制NTK和深神经网络的内核回归之间存在性能差距。 我们研究了有限宽度的神经网络的动态,并导出了微分方程的无限层次,神经切线层级(第n)。 我们证明,在P级别P≥2处截断的第n个层次结构近似于在神经网络宽度和数据设定维度上的某些条件下对NTK的动态达到任意精度。 随着P增加,这些近似所需的假设变弱。 最后,NTH可以被视为NTK的更高阶扩展。 特别地,在P = 2处截断的Nth截止了NTK动态。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号