2015 IEEE International Congress on Big Data 2015
DOI: 10.1109/bigdatacongress.2015.64
|View full text |Cite
|
Sign up to set email alerts
|

Machine Learning-Based Configuration Parameter Tuning on Hadoop System

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 40 publications
(13 citation statements)
references
References 16 publications
0
13
0
Order By: Relevance
“…Machine learning techniques have been applied to explore complex configuration spaces to find near optimal settings without considering constraints on operating behavior [5,53,60,68]. Some approaches employ ML to meet resource constraints in dynamic environments [9].…”
Section: Motivationmentioning
confidence: 99%
See 1 more Smart Citation
“…Machine learning techniques have been applied to explore complex configuration spaces to find near optimal settings without considering constraints on operating behavior [5,53,60,68]. Some approaches employ ML to meet resource constraints in dynamic environments [9].…”
Section: Motivationmentioning
confidence: 99%
“…Machine learning frameworks Many learning approaches have been proposed for predicting an optimal configuration within a complicated configuration space [8-10, 31, 40, 53, 65]. Machine learning has even been applied to further improve existing heuristic autotuners, like Starfish [21], by using learning models to direct the search for optimal configurations [5,60]. Perhaps the most closely related learning works are those based on reinforcement learning (RL) [51].…”
Section: Related Workmentioning
confidence: 99%
“…Another work [25] proposes a tree-based regression approach consisting of a prediction and an optimization phase. The former one estimates the execution time of a MapReduce job by building three prediction models.…”
Section: Batch Processing Systemsmentioning
confidence: 99%
“…For the experiments, we will present our cluster performance based on MapReduce and Spark using the HiBench suite [23,23]. In particular, we have selected two Hibench workloads out of thirteen standard workloads to represent the two types of jobs, namely WordCount (aggregation job) [32], and TeraSort (shuffle job) [33] with large datasets. We selected both the workloads because of their complex characteristics to study how efficiently both the workloads analyze the cluster performance by correlating MapReduce and Spark function with a combination of groups of parameters.…”
Section: Cluster Architecturementioning
confidence: 99%