Proceedings of the Tenth European Conference on Computer Systems 2015
DOI: 10.1145/2741948.2741952
|View full text |Cite
|
Sign up to set email alerts
|

Deriving and comparing deduplication techniques using a model-based classification

Abstract: Data deduplication has been a hot research topic and a large number of systems have been developed. These systems are usually seen as an inherently linked set of characteristics.However, a detailed analysis shows independent concepts that can be used in other systems.In this work, we perform this analysis on the main representatives of deduplication systems. We embed the results in a model, which shows two yet unexplored combinations of characteristics. In addition, the model enables a comprehensive evaluation… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2016
2016
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 23 publications
(25 reference statements)
0
2
0
Order By: Relevance
“…A feature of backup workloads is strong locality, which has been used to tackle the challenges facing deduplication-based backup storage. [27][28][29] Here locality refers to the fact that current backup stream tend to have patterns that correspond to an earlier backup stream. For example, in order to accelerate the process of detecting duplicate chunks, the system groups consecutive unique chunks into fixed-sized containers to preserve their F I G U R E 1 An example of data deduplication.…”
Section: Background Of Data Reduction Techniquesmentioning
confidence: 99%
“…A feature of backup workloads is strong locality, which has been used to tackle the challenges facing deduplication-based backup storage. [27][28][29] Here locality refers to the fact that current backup stream tend to have patterns that correspond to an earlier backup stream. For example, in order to accelerate the process of detecting duplicate chunks, the system groups consecutive unique chunks into fixed-sized containers to preserve their F I G U R E 1 An example of data deduplication.…”
Section: Background Of Data Reduction Techniquesmentioning
confidence: 99%
“…However, we vary the number of used processes in Section V-C. Table I shows the different sizes of the checkpoints. c) Deduplication: We analyzed each checkpoint with the FS-C deduplication tool suite [49], which has already been applied in several deduplication studies [50], [51]. We chose fixed-sized chunking and content-defined chunking (CDC) as chunking methods.…”
Section: Deduplication Of Checkpointsmentioning
confidence: 99%