首页> 外文期刊>Computational Intelligence and AI in Games, IEEE Transactions on >Preference Learning for Move Prediction and Evaluation Function Approximation in Othello
【24h】

Preference Learning for Move Prediction and Evaluation Function Approximation in Othello

机译:Othello中的移动预测和评估函数逼近的偏好学习

获取原文
获取原文并翻译 | 示例
           

摘要

This paper investigates the use of preference learning as an approach to move prediction and evaluation function approximation, using the game of Othello as a test domain. Using the same sets of features, we compare our approach with least squares temporal difference learning, direct classification, and with the Bradley–Terry model, fitted using minorization–maximization (MM). The results show that the exact way in which preference learning is applied is critical to achieving high performance. Best results were obtained using a combination of board inversion and pair-wise preference learning. This combination significantly outperformed the others under test, both in terms of move prediction accuracy, and in the level of play achieved when using the learned evaluation function as a move selector during game play.
机译:本文以奥赛罗博弈为测试域,研究了偏好学习作为移动预测和评估函数逼近的一种方法。使用相同的功能集,我们将我们的方法与最小二乘时差学习,直接分类以及使用最小化-最大化(MM)拟合的Bradley-Terry模型进行比较。结果表明,应用偏好学习的确切方法对于实现高性能至关重要。结合电路板反转和成对偏好学习可获得最佳结果。无论是在移动预测准确性方面,还是在游戏过程中使用学习的评估功能作为移动选择器时,这种组合在性能上均明显优于其他测试对象。<​​/ p>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号