首页> 外文会议>IEEE Annual Consumer Communications and Networking Conference >qMDP: DASH Adaptation using Queueing Theory within a Markov Decision Process
【24h】

qMDP: DASH Adaptation using Queueing Theory within a Markov Decision Process

机译:QMDP:在马尔可夫决策过程中使用排队理论进行延线适应

获取原文

摘要

Adaptive bitrate (ABR) streaming algorithms play an important role in ensuring a high Quality of Experience (QoE) for the consumer. However, a lot of ABR algorithms tend to be too ad hoc. In response, methods based on a Markov Decision Process (MDP) offer more intelligent models. In particular, Reinforcement Learning (RL) methods typically do so via QoE metrics. However, RL methods are plagued by high complexity and long convergence times due to their model-free nature. This paper proposes qMDP, which is an RL method with an MDP partially modeled by an M/D/1/K queue. Our study shows that qMDP results in higher QoE and faster convergence compared to a QoE-only model-free version.
机译:自适应比特率(ABR)流媒体算法在确保消费者的高质量经验(QoE)方面发挥着重要作用。然而,很多ABR算法往往是太临时。作为响应,基于Markov决策过程的方法(MDP)提供了更智能的模型。特别地,加强学习(RL)方法通常通过QoE度量来实现。然而,由于无模型性质,RL方法血液高复杂性和长收敛时间困扰。本文提出了QMDP,其是具有由M / D / 1 / k队列部分建模的MDP的RL方法。我们的研究表明,与QoE的无模型版本相比,QMDP会导致更高的QoE和更快的融合。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号