首页> 外文会议>Proceedings of the 9th international conference for young computer scientists (ICYCS 2008) >Influence Graph based Task Decomposition and State Abstraction in Reinforcement Learning
【24h】

Influence Graph based Task Decomposition and State Abstraction in Reinforcement Learning

机译:强化学习中基于影响图的任务分解和状态抽象

获取原文

摘要

Task decomposition and State abstraction are crucial parts in reinforcement learning. It allows an agent to ignore aspects of its current states that are irrelevant to its current decision, and therefore speeds up dynamic programming and learning. This paper presents the SVI algorithm that uses a dynamic Bayesian network model to construct an influence graph that indicates relationships between state variables. SVI performs state abstraction for each subtask by ignoring irrelevant state variables and lower level subtasks. Experiment results show that the decomposition of tasks introduced by SVI can significantly accelerate constructing a near-optimal policy. This general framework can be applied to a broad spectrum of complex real world problems such as robotics, industrial manufacturing, games and others.
机译:任务分解和状态抽象是强化学习的关键部分。它允许代理忽略与其当前决定无关的其当前状态的各个方面,从而加快了动态编程和学习的速度。本文提出了使用动态贝叶斯网络模型构建表示状态变量之间关系的影响图的SVI算法。 SVI通过忽略无关的状态变量和较低级别的子任务来为每个子任务执行状态抽象。实验结果表明,SVI引入的任务分解可以显着加速构建近乎最优的策略。该通用框架可应用于各种复杂的现实世界问题,例如机器人技术,工业制造,游戏等。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号