首页> 外文会议>International Conference on Robot Intelligence Technology and Applications >An Empirical Study on the Optimal Batch Size for the Deep Q-Network
【24h】

An Empirical Study on the Optimal Batch Size for the Deep Q-Network

机译:深度Q网的最佳批量大小的实证研究

获取原文

摘要

We empirically find the optimal batch size for training the Deep Q-network on the cart-pole system. The efficiency of the training is evaluated by the performance of the network on task after training, and total time and steps required to train. The neural network is trained for 10 different sizes of batch with other hyper parameter values fixed. The network is able to carry out the cart-pole task with the probability of 0.99 or more with the batch sizes from 8 to 2048. The training time per step for training tends to increase linearly, and the total steps for training decreases more than exponentially as the batch size increases. Due to these tendencies, we could empirically observe the quadratic relationship between the total time for training and the logarithm of batch size, which is convex, and the optimal batch size that minimizes training time could also be found. The total steps and time for training are minimum at the batch size 64. This result can be expanded to other learning algorithm or tasks, and further, theoretical analysis on the relationship between the size of batch or other hyper-parameters and the efficiency of training from the optimization point of view.
机译:我们经验验证找到最佳批量尺寸,用于训练推车系统上的深Q网络。培训的效率是通过网络在培训后的任务上的性能进行评估,以及培训所需的总时间和步骤。神经网络培训,有10种不同大小的批次,具有固定的其他超参数值。网络能够从8到2048的批量尺寸执行卡车杆任务,批量尺寸为0.99或更多。每步训练训练趋于线性地增加,并且训练的总步骤比指数更低随着批量尺寸的增加。由于这些趋势,我们可以经验遵守训练总时间与批次尺寸的对数之间的二次关系,也可以找到最小化训练时间的最佳批量尺寸。批量大小的培训总步长和时间最小。该结果可以扩展到其他学习算法或任务,以及对批量大小或其他超参数之间关系的理论分析以及培训效率之间的关系。从优化的角度来看。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号