2019
DOI: 10.1587/transinf.2018edp7297
|View full text |Cite
|
Sign up to set email alerts
|

High-Performance End-to-End Integrity Verification on Big Data Transfer

Abstract: The scale of scientific data generated by experimental facilities and simulations in high-performance computing facilities has been proliferating with the emergence of IoT-based big data. In many cases, this data must be transmitted rapidly and reliably to remote facilities for storage, analysis, or sharing, for the Internet of Things (IoT) applications. Simultaneously, IoT data can be verified using a checksum after the data has been written to the disk at the destination to ensure its integrity. However, thi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 34 publications
0
3
0
Order By: Relevance
“…The primary objective of big data transfer frameworks is to ensure the accuracy of transferred data. To accomplish this, extensive research has been conducted on the design and optimization of data integrity verification systems in different contexts, including storage systems [29,30], cloud-based storage [31,32], file systems [33,34], and data transfer systems [35][36][37][38].…”
Section: Data Integritymentioning
confidence: 99%
See 1 more Smart Citation
“…The primary objective of big data transfer frameworks is to ensure the accuracy of transferred data. To accomplish this, extensive research has been conducted on the design and optimization of data integrity verification systems in different contexts, including storage systems [29,30], cloud-based storage [31,32], file systems [33,34], and data transfer systems [35][36][37][38].…”
Section: Data Integritymentioning
confidence: 99%
“…For this reason, a parallel checksum approach is necessary and preferable. Many researchers have proposed methods for transferring multiple files in parallel by exploiting pipelining and parallelism [35][36][37]39] techniques. Unlike serial and file-based data transfer tools, this work considers transferring the workload as objects rather than files.…”
Section: Data Integritymentioning
confidence: 99%
“…Ensuring the correctness of the transferred data is one of the primary concerns of data transfer frameworks. Many studies have been performed on the design and implementation of data integrity verification and their optimization in data storage [34,36,37], cloud storage [38][39][40][41][42], file systems [32][33][34][35]43], databases [32,44,45], and data transfer systems [18,19,46,47].…”
Section: Related Workmentioning
confidence: 99%
“…End-to-end data integrity verification in large scale data transfers is not only essential but also very expensive. It increases the amount of disk I/O and processing needed for the data transfer, thereby reducing the overall data transfer performance [17,18]. In this work, our aim is to design a data integrity verification framework for object-based big data transfer systems.…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, modern Internet application software systems have built-in measurement and monitoring capabilities to ensure the application performance and mitigate the effects of failures in the underlying network system layer [7], [8], [9]. From the gray failure diagnosis perspective, the fundamental ML based solution approach appears to be a very viable choice with its promise in learning a model to map the observations to internal faulty behaviors of the network.…”
Section: Introductionmentioning
confidence: 99%