2016
DOI: 10.1007/978-3-319-44881-7_11
|View full text |Cite
|
Sign up to set email alerts
|

Fault Tolerance in MapReduce: A Survey

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 14 publications
(3 citation statements)
references
References 28 publications
0
3
0
Order By: Relevance
“…Scientific applications are increasingly implemented to tolerate faults [3,26,20,13,17]. The three main techniques for implementing fault-tolerant algorithms are Algorithm-Based Fault-Tolerance [32,9], restarting failed sub-jobs [23], and checkpointing/restart [18,17]. Checkpointing libraries can save their checkpoint either to a (possibly network attached) disk or to the compute node's main memory ("diskless") [27].…”
Section: Related Workmentioning
confidence: 99%
“…Scientific applications are increasingly implemented to tolerate faults [3,26,20,13,17]. The three main techniques for implementing fault-tolerant algorithms are Algorithm-Based Fault-Tolerance [32,9], restarting failed sub-jobs [23], and checkpointing/restart [18,17]. Checkpointing libraries can save their checkpoint either to a (possibly network attached) disk or to the compute node's main memory ("diskless") [27].…”
Section: Related Workmentioning
confidence: 99%
“…Replication has been successfully employed and practiced to ensure high data availability in large-scale distributed storage systems [17,27,35]. Moreover, replication can be leveraged to improve data access performance under high load.…”
Section: Introductionmentioning
confidence: 99%
“…The complexity, heterogeneity, dynamism and inherently distributed nature of Big Data technologies do not help either for this purpose. Even models enjoying a straightforward adaptability to Big Data computing environments (e.g., ensembles for predictive modeling) can be severely affected by the obsolescence of the information from where they are learned [ 3 ], or the failure of a node in a distributed Map-Reduce computing grid [ 4 ]. All in all, data fusion, processing, learning and visualization of Big Data require a major focus not only on tailoring the algorithmic steps underlying each model/technique to the computing technologies underneath, but also endowing them with higher levels of resilience against failures, adaptation to changes in data and the accommodation of unprecedented levels of data volume, heterogeneity and veracity.…”
Section: Introductionmentioning
confidence: 99%