2019
DOI: 10.1145/3328740
|View full text |Cite
|
Sign up to set email alerts
|

A New Framework for Evaluating Straggler Detection Mechanisms in MapReduce

Abstract: Big Data systems (e.g., Google MapReduce, Apache Hadoop, Apache Spark) rely increasingly on speculative execution to mask slow tasks, also known as stragglers, because a job's execution time is dominated by the slowest task instance. Big Data systems typically identify stragglers and speculatively run copies of those tasks with the expectation that a copy may complete faster to shorten job execution times. There is a rich body of recent results on straggler mitigation in MapReduce. However, the majority of the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
16
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(16 citation statements)
references
References 18 publications
0
16
0
Order By: Relevance
“…In order to accelerate learning, AdaSGD relies on a system parameter: the expected percentage of nonstragglers (denoted by s%). We highlight that this value is not a hyperparameter that needs tuning for each ML application but a system parameter that solely depends on the computing and networking characteristics of the workers, while it can be adapted dynamically [64,65]. We define the staleness-aware dampening factor Λ(τ ) = e −βτ , with β chosen s.t.…”
Section: Adaptive Stochastic Gradient Descentmentioning
confidence: 99%
“…In order to accelerate learning, AdaSGD relies on a system parameter: the expected percentage of nonstragglers (denoted by s%). We highlight that this value is not a hyperparameter that needs tuning for each ML application but a system parameter that solely depends on the computing and networking characteristics of the workers, while it can be adapted dynamically [64,65]. We define the staleness-aware dampening factor Λ(τ ) = e −βτ , with β chosen s.t.…”
Section: Adaptive Stochastic Gradient Descentmentioning
confidence: 99%
“…The framework was evaluated within an 8-node Hadoop cluster, whereby they injected stragglers to alleviate job data skew. Tien-Dat et al [16] proposed a framework for straggler detection and mitigation to enhance job execution time and system energyefficiency. Using the Grid'5000 testbed consisting of 21-nodes, authors artificially injected stragglers into job application execution and evaluated the framework with straggler mitigation techniques.…”
Section: Related Workmentioning
confidence: 99%
“…Conducting such experiments allows for empirical study of realistic system operation in order to propose new approaches without interfering with production system behavior, as well as underpin parameterization of Cloud simulation frameworks [9]. This is particularly important as current simulators are unable to realistically represent straggler manifestation due to their decoupling of occurrence probability and the underlying cluster operational conditions [14][15][16].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Related Works: The problem of tasks duration variability has received significant attention in the context of distributed computing, first in Grid platforms [19] and later in Cloud computing platforms in the context of MapReduce environment [20]. Some of the work in this area consists one one hand in analyzing the variability of task execution times and trying to explain it [21,22]. On the other hand, several strategies have been designed to mitigate the effect of variability.…”
Section: Introductionmentioning
confidence: 99%