首页> 外文会议>International Conference on Computer Communication and Networks >A Comparison of Distributed Machine Learning Platforms
【24h】

A Comparison of Distributed Machine Learning Platforms

机译:分布式机器学习平台的比较

获取原文

摘要

The proliferation of big data and big computing boosted the adoption of machine learning across many application domains. Several distributed machine learning platforms emerged recently. We investigate the architectural design of these distributed machine learning platforms, as the design decisions inevitably affect the performance, scalability, and availability of those platforms. We study Spark as a representative dataflow system, PMLS as a parameter-server system, and TensorFlow and MXNet as examples of more advanced dataflow systems. We take a distributed systems perspective, and analyze the communication and control bottlenecks for these approaches. We also consider fault-tolerance and ease-of-development in these platforms. In order to provide a quantitative evaluation, we evaluate the performance of these three systems with basic machine learning tasks: logistic regression, and an image classification example on the MNIST dataset.
机译:大数据和大计算的扩散提高了在许多应用领域的机器学习的采用。最近出现了几种分布式机器学习平台。我们调查了这些分布式机器学习平台的架构设计,因为设计决策不可避免地影响这些平台的性能,可扩展性和可用性。我们将Spark作为代表性数据流系统,PMLS作为参数服务器系统,以及TensorFlow和MXNet作为更高级数据流系统的示例。我们采用分布式系统的透视图,分析了这些方法的通信和控制瓶颈。我们还考虑这些平台的容错和易于开发。为了提供定量评估,我们评估这三个系统的性能与基本机器学习任务:Logistic回归和Mnist DataSet上的图像分类示例。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号