首页> 外文会议>Brazilian Conference on Intelligent Systems >Object-Oriented Reinforcement Learning in Cooperative Multiagent Domains
【24h】

Object-Oriented Reinforcement Learning in Cooperative Multiagent Domains

机译:协作多主体领域中的面向对象的强化学习

获取原文

摘要

Although Reinforcement Learning methods have successfully been applied to increasingly large problems, scalability remains a central issue. While Object-Oriented Markov Decision Processes (OO-MDP) are used to exploit regularities in a domain, Multiagent System (MAS) methods are used to divide workload amongst multiple agents. In this work we propose a novel combination of OO-MDP and MAS, called Multiagent Object-Oriented Markov Decision Process (MOO-MDP), so as to accrue the benefits of both strategies and be able to better address scalability issues. We present an algorithm to solve deterministic cooperative MOO-MDPs, and prove that it learns optimal policies while reducing the learning space by exploiting state abstractions. We experimentally compare our results with earlier approaches and show advantages with regard to discounted cumulative reward, number of steps to fulfill the task, and Q-table size.
机译:尽管强化学习方法已成功应用于越来越大的问题,但可扩展性仍然是中心问题。虽然使用面向对象的马尔可夫决策过程(OO-MDP)来利用域中的规则性,但是使用多代理系统(MAS)方法来在多个代理之间分配工作量。在这项工作中,我们提出了OO-MDP和MAS的新颖组合,称为多代理面向对象的马尔可夫决策过程(MOO-MDP),以便从这两种策略中受益,并能够更好地解决可伸缩性问题。我们提出了一种解决确定性合作MOO-MDP的算法,并证明了它在学习最优策略的同时通过利用状态抽象来减少学习空间。我们通过实验将我们的结果与早期方法进行比较,并显示出在折扣累积奖励,完成任务的步骤数以及Q表大小方面的优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号