2015 International Conference on Pervasive Computing (ICPC) 2015
DOI: 10.1109/pervasive.2015.7087116
|View full text |Cite
|
Sign up to set email alerts
|

A survey and comparative study of data deduplication techniques

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 21 publications
(5 citation statements)
references
References 5 publications
0
5
0
Order By: Relevance
“…The size/length of the output of a hash function does not depend on the length of the input. Hash can be regarded as a ”signature” for a given text [ 49 , 50 , 51 ]. One of the major applications of hash functions lies in the field of multimedia broadcast networks, as a content identifier [ 49 , 50 , 51 ].…”
Section: Proposed Distributed Architecture and Clustering Algorithmentioning
confidence: 99%
See 1 more Smart Citation
“…The size/length of the output of a hash function does not depend on the length of the input. Hash can be regarded as a ”signature” for a given text [ 49 , 50 , 51 ]. One of the major applications of hash functions lies in the field of multimedia broadcast networks, as a content identifier [ 49 , 50 , 51 ].…”
Section: Proposed Distributed Architecture and Clustering Algorithmentioning
confidence: 99%
“…Hash can be regarded as a ”signature” for a given text [ 49 , 50 , 51 ]. One of the major applications of hash functions lies in the field of multimedia broadcast networks, as a content identifier [ 49 , 50 , 51 ]. The hash function aids the network by providing content identification to easily determine which content has been broadcasted, timing information, and to what station.…”
Section: Proposed Distributed Architecture and Clustering Algorithmentioning
confidence: 99%
“…Although integrating data deduplication with file migration can improve the slow tier space utilization and potentially reduce migration cost, it also brings some performance issues such as high compute and memory resource utilization, high latency, and low throughput [29][30][31]. The chunking process, chunk ID generation, and chunk ID searches in the indexing table are timeconsuming.…”
Section: Deduplication and Challenges Of Integrating Deduplication Wimentioning
confidence: 99%
“…However, a full content based hashing calculation may increase high computation cost [3]. A compromised way is taking a partial content based hashing calculation, which may bring a faster response to users, with a few sacrifices of deduplication inaccuracy [4,5]. This work in the paper aims at how to design and implement the various file deduplication schemes for space saving.…”
Section: Introductionmentioning
confidence: 99%