首页> 外文会议>International Conference on Artificial Neural Networks >Hierarchical Reinforcement Learning with Unlimited Recursive Subroutine Calls
【24h】

Hierarchical Reinforcement Learning with Unlimited Recursive Subroutine Calls

机译:无限递归子例程调用的分层强化学习

获取原文

摘要

Humans can set suitable subgoals to achieve certain tasks. They can also set sub-subgoals recursively if required. The depth of this recursion is apparently unlimited. Inspired by this behavior, we propose a new hierarchical reinforcement learning architecture called RGoal. RGoal solves the Markov Decision Process (MDP) in an augmented state-action space. In multitask settings, sharing subroutines between tasks makes learning faster. A novel mechanism called thought-mode is a type of model-based reinforcement learning. It combines learned simple tasks to solve unknown complicated tasks rapidly, sometimes in zero-shot time.
机译:人类可以设置合适的子目标来完成某些任务。如果需要,他们还可以递归设置子子目标。递归的深度显然是无限的。受此行为的启发,我们提出了一种称为RGoal的新的分层强化学习架构。 RGoal在增强的状态行动空间中解决了马尔可夫决策过程(MDP)。在多任务设置中,在任务之间共享子例程可以使学习更快。一种称为思维模式的新颖机制是一种基于模型的强化学习。它结合了学到的简单任务,可以快速解决未知的复杂任务,有时甚至可以零触发。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号