首页> 外文会议>International Conference on Artificial Neural Networks >Improving Weight Initialization of ReLU and Output Layers
【24h】

Improving Weight Initialization of ReLU and Output Layers

机译:改善ReLU和输出层的权重初始化

获取原文

摘要

We introduce a data-dependent weight initialization scheme for ReLU and output layers commonly found in modern neural network architectures. An initial feedforward pass through the network is performed using an initialization set (a subset of the training data set). Using statistics obtained from this pass, we initialize the weights of the network, so the following properties are met: (1) weight matrices are orthogonal; (2) ReLU layers produce a predetermined fraction of nonzero activations; (3) the outputs produced by internal layers have a predetermined variance; (4) weights in the last layer are chosen to minimize the squared error in the initialization set. We evaluate our method on popular architectures (VGG16, VGG19, and InceptionV3) and faster convergence rates are achieved on the ImageNet data set when compared to state-of-the-art initialization techniques (LSUV, He, and Glorot).
机译:我们为ReLU和现代神经网络体系结构中常见的输出层引入了一种与数据相关的权重初始化方案。使用初始化集(训练数据集的子集)执行通过网络的初始前馈。使用从此遍获得的统计信息,我们初始化网络的权重,因此可以满足以下属性:(1)权重矩阵是正交的; (2)ReLU层产生预定分数的非零激活; (3)内部层产生的输出具有预定的方差; (4)选择最后一层的权重以最小化初始化集中的平方误差。我们在流行的架构(VGG16,VGG19和InceptionV3)上评估了我们的方法,并且与最新的初始化技术(LSUV,He和Glorot)相比,在ImageNet数据集上实现了更快的收敛速度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号