首页> 美国政府科技报告 >Countable State Discounted Markovian Decision Processes with Unbounded Rewards
【24h】

Countable State Discounted Markovian Decision Processes with Unbounded Rewards

机译:具有无限奖励的可数州折现马尔可夫决策过程

获取原文

摘要

Countable state, finite action Markov decision processes are investigated under a criterion of maximizing expected discounted rewards over an infinite planning horizon. Well-known results of Maitra and Blackwell are generalized, their assumption of bounded rewards being replaced by the following weaker condition: the expected absolute reward to be received at time n+1 minus the actual absolute reward received at time n (as a function of the state of the system of time n, the action taken at time n, and the decision rule to be followed at time n+1) can be bounded above. Under this condition it is shown that the expected discounted reward (over the infinite planning horizon) from each policy is finite and that there exists a stationary policy which is optimal. Additional results are presented concerning the policy improvement and successive approximations algorithms for computation of optimal policies. All of these results are extended to Markov renewal decision processes under one additional condition on the transition time distributions. As in Blackwell's work on discounted dynamic programming a central role is played by Banach's fixed point theorem for contraction mappings. Examples are presented of inventory and queueing control problems which satisfy our assumptions but do not exhibit bounded rewards. (Author)

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号