首页> 外文期刊>AI communications >Deterministic policies based on maximum regrets in MDPs with imprecise rewards
【24h】

Deterministic policies based on maximum regrets in MDPs with imprecise rewards

机译:Deterministic policies based on maximum regrets in MDPs with imprecise rewards

获取原文
获取原文并翻译 | 示例
           

摘要

Markov Decision Process Models (MDPs) are a powerful tool for planning tasks and sequential decision-making issues. In this work we deal with MDPs with imprecise rewards, often used when dealing with situations where the data is uncertain. In this context, we provide algorithms for finding the policy that minimizes the maximum regret. To the best of our knowledge, all the regret-based methods proposed in the literature focus on providing an optimal stochastic policy. We introduce for the first time a method to calculate an optimal deterministic policy using optimization approaches. Deterministic policies are easily interpretable for users because for a given state they provide a unique choice. To better motivate the use of an exact procedure for finding a deterministic policy, we show some (theoretical and experimental) cases where the intuitive idea of using a deterministic policy obtained after "determinizing" the optimal stochastic policy leads to a policy far from the exact deterministic policy.

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号