2013 IEEE International Symposium on Workload Characterization (IISWC) 2013
DOI: 10.1109/iiswc.2013.6704674
|View full text |Cite
|
Sign up to set email alerts
|

Characterizing the efficiency of data deduplication for big data storage management

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 37 publications
(16 citation statements)
references
References 6 publications
0
15
0
Order By: Relevance
“…On ZFS file system [42] with inline deduplication, the average CPU usage could boost from 20% to 60% due to heavy computation and indexing. The inline deduplication causes the storage system to consume 10% more power on average [44].…”
Section: Backgroundsmentioning
confidence: 99%
“…On ZFS file system [42] with inline deduplication, the average CPU usage could boost from 20% to 60% due to heavy computation and indexing. The inline deduplication causes the storage system to consume 10% more power on average [44].…”
Section: Backgroundsmentioning
confidence: 99%
“…Three main reasons for data redundancy are: (1) addition of nodes, (2) expansion of datasets, and (3) data replication. The addition of a single virtual machine (VM) brings around 97% more redundancy, and the growth in large datasets comes with 47% redundant data points [13]. In addition, the storage mechanism for maximum data availability (also called data replication) brings 100% redundancy at the cluster level.…”
Section: Data Deduplication (Redundancy Elimination)mentioning
confidence: 99%
“…In contrast, global deduplication can achieve maximum redundancy but compromises on the hashing overheads. In addition, finegrained deduplication is not suitable for big datasets especially in streaming data environments [13].…”
Section: Open Research Issuesmentioning
confidence: 99%
See 2 more Smart Citations