2019
DOI: 10.1109/access.2019.2926195
|View full text |Cite
|
Sign up to set email alerts
|

MII: A Novel Content Defined Chunking Algorithm for Finding Incremental Data in Data Synchronization

Abstract: In the data backup system, to reduce the bandwidth and processing time overhead caused by full backup technology during data synchronization between backups and source data, incremental backup technology is emerging as the focus of academic and industrial research. It is key but poorly-solved to find the incremental data between backups and source data for incremental backup technology. To find out the incremental data during the backup process, here, in this paper, we propose a novel content-defined chunking … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 48 publications
0
3
0
Order By: Relevance
“…Low entropy strings are strings which consist of repetitive bytes or patterns. This challenge means it is preferable for the algorithm to be able to eliminate the redundancy within this kind of string [32]. 4) High throughput [33].…”
Section: B Motivationmentioning
confidence: 99%
See 1 more Smart Citation
“…Low entropy strings are strings which consist of repetitive bytes or patterns. This challenge means it is preferable for the algorithm to be able to eliminate the redundancy within this kind of string [32]. 4) High throughput [33].…”
Section: B Motivationmentioning
confidence: 99%
“…In the previous research of our team, MII algorithm was proposed to achieve better ability of resistance against the byte shifting by sacrificing the stability of chunk size [32]. The pseudo code and chunking process of MII algorithm are shown in Fig.…”
Section: B Motivationmentioning
confidence: 99%
“…In the same way, the block level searches the content of the block, eliminates one copy of the league, and retains another block. The block-level of the file involves four processing steps; chunking, fingerprinting, indexing of fingerprints, and managing the stored information of data [5,6].…”
Section: Introductionmentioning
confidence: 99%
“…In fixed-length block deduplication, data is divided into chunks of a constant size, whereas in variable-length block deduplication, data is divided into distinct chunks based on different factors [9,12]. While block-level deduplication is more efficient than filelevel Deduplication, it requires more system resources [13,14].…”
Section: Introductionmentioning
confidence: 99%