首页> 外文会议>International Conference on Robot Intelligence Technology and Applications >An Empirical Study on the Optimal Batch Size for the Deep Q-Network
【24h】

An Empirical Study on the Optimal Batch Size for the Deep Q-Network

机译:深度Q网的最佳批量大小的实证研究

获取原文

摘要

We empirically find the optimal batch size for training the Deep Q-network on the cart-pole system. The efficiency of the training is evaluated by the performance of the network on task after training, and total time and steps required to train. The neural network is trained for 10 different sizes of batch with other hyper parameter values fixed. The network is able to carry out the cart-pole task with the probability of 0.99 or more with the batch sizes from 8 to 2048. The training time per step for training tends to increase linearly, and the total steps for training decreases more than exponentially as the batch size increases. Due to these tendencies, we could empirically observe the quadratic relationship between the total time for training and the logarithm of batch size, which is convex, and the optimal batch size that minimizes training time could also be found. The total steps and time for training are minimum at the batch size 64. This result can be expanded to other learning algorithm or tasks, and further, theoretical analysis on the relationship between the size of batch or other hyper-parameters and the efficiency of training from the optimization point of view.
机译:我们凭经验找到最佳的批量大小的车极系统的培训深Q-网络。培训的效率是通过对任务的网络训练后的性能评估,总时间和步骤所需的培训。神经网络进行训练,对于10点不同批次的尺寸与固定其他超参数值。网络能够与0.99或更多个与所述批量大小从8的概率车极任务进行到2048。训练时间每步训练趋于线性增加,以及用于训练的总步骤减小多于呈指数如批量大小增加而增加。由于这些趋势,我们可以观察经验的总培训时间和批量尺寸的对,这是凸的,和最佳批量大小减少了培训的时间也可能被发现之间的二次关系。用于训练的总的步骤和时间是最小的批量大小64该结果可以被扩展到其它学习算法或任务,并且进一步地,在分批或其他超参数的大小和训练的效率之间的关系的理论分析但从优化点。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号