首页> 外文会议>International Conference on Engineering Technology >A cloud based model for deduplication of large data
【24h】

A cloud based model for deduplication of large data

机译:基于云的大数据重复数据删除模型

获取原文

摘要

These days the amount of data that is generated or stored in any system is so high and the amount of storage required is large. By consuming a large number of storage space, the processing speed and the retrieval speed of the data is also slow. Deduplication is the process of identifying redundant data from a large amount of given data before storing them to any physical storage. This way the storage space consumed can be reduced. The concept of deduplication is used in many areas and it is one of the challenging research topics in recent times. In this paper, a cloud based model is proposed for implementing deduplication of a large amount of data available. The model comprises of both the deduplication of data before uploading to the cloud storage and the reverse deduplication of data when downloading the necessary data. The paper also describes the various issues faced during deduplication and the different data structures and algorithms that can be used for effective deduplication.
机译:如今,在任何系统中生成或存储的数据量如此之高,所需的存储量很大。通过消耗大量的存储空间,处理速度和数据的检索速度也很慢。重复数据删除是在将它们存储到任何物理存储之前从大量给定数据识别冗余数据的过程。这样可以减少消耗的存储空间。重复数据删除的概念用于许多领域,它是最近的挑战性研究主题之一。在本文中,提出了一种基于云的模型,用于实施大量数据的重复数据删除。该模型包括在上传到云存储之前数据的重复数据删除以及下载必要数据时数据的反向重复数据删除。本文还描述了重复数据删除期间面临的各种问题以及可用于有效重复数据删除的不同数据结构和算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号