首页> 外文会议>Intelligent Control, 1991., Proceedings of the 1991 IEEE International Symposium on >A comparison of supervised and reinforcement learning methods on a reinforcement learning task
【24h】

A comparison of supervised and reinforcement learning methods on a reinforcement learning task

机译:强化学习任务的监督学习和强化学习方法比较

获取原文

摘要

The forward modeling approach of M.I. Jordan and J.E. Rumelhart (1990) has been shown to be applicable when supervised learning methods are to be used for solving reinforcement learning tasks. Because such tasks are natural candidates for the application of reinforcement learning methods, there is a need to evaluate the relative merits of these two learning methods on reinforcement learning tasks. The author presents one such comparison on a task involving learning to control an unstable, nonminimum phase, dynamic system. The comparison shows that the reinforcement learning method used performs better than the supervised learning method. An examination of the learning behavior of the two methods indicates that the differences in performance can be attributed to the underlying mechanics of the two learning methods, which provides grounds for believing that similar performance differences can be expected on other reinforcement learning tasks as well.
机译:M.I.的正向建模方法乔丹和鲁默哈特(J.E. Rumelhart)(1990)已证明适用于将监督学习方法用于解决强化学习任务的情况。由于此类任务是强化学习方法应用的自然候选者,因此需要评估这两种学习方法在强化学习任务上的相对优点。作者对一项涉及学习控制不稳定,非最小相位,动态系统的任务进行了这样的比较。比较表明,所使用的强化学习方法比监督学习方法表现更好。对这两种方法的学习行为的检查表明,性能的差异可以归因于两种学习方法的内在机理,这为相信在其他强化学习任务上也预期会有类似的性能差异提供了依据。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号