2009
DOI: 10.1016/j.parco.2009.07.001
|View full text |Cite
|
Sign up to set email alerts
|

Tuning parallel applications in parallel

Abstract: Auto-tuning has recently received significant attention from the High Performance Computing community. Most auto-tuning approaches are specialized to work either on specific domains such as dense linear algebra and stencil computations, or only at certain stages of program execution such as compile time and runtime.Real scientific applications, however, demand a cohesive environment that can efficiently provide auto-tuning solutions at all stages of application development and deployment. Towards that end, we … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2011
2011
2019
2019

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 66 publications
0
8
0
Order By: Relevance
“…In this work multiple local searches are used to find the estimate of the global optimum independently. This is in contrast to the most popular hybrid global optimisation methods, where local search is used only to refine x * in the regions suggested by the initial stage of the global search [24][25][26]. While a hybrid scheme can reduce the required number of starting points and thus the number of evaluations of , it necessarily introduces latency because the choice of new starting points is made after x * has been evaluated from some previous starting points.…”
Section: Modelling and Simulation In Engineeringmentioning
confidence: 99%
“…In this work multiple local searches are used to find the estimate of the global optimum independently. This is in contrast to the most popular hybrid global optimisation methods, where local search is used only to refine x * in the regions suggested by the initial stage of the global search [24][25][26]. While a hybrid scheme can reduce the required number of starting points and thus the number of evaluations of , it necessarily introduces latency because the choice of new starting points is made after x * has been evaluated from some previous starting points.…”
Section: Modelling and Simulation In Engineeringmentioning
confidence: 99%
“…For example, MATE was able to scale up to 32 cores as shown in Caymes-Scutari et al 16 and Morajko et al, 17 while Active Harmony scaled up to 128 cores as described in Tiwari et al 18 …”
Section: Validation Of Proposed Model For Hierarchical Dynamic Tuningmentioning
confidence: 99%
“…Moreover, it is worth noticing the huge improvement that the ELASTIC approach represents in comparison to previous centralized dynamic tuning tools. For example, MATE was able to scale up to 32 cores as shown in Caymes‐Scutari et al and Morajko et al, while Active Harmony scaled up to 128 cores as described in Tiwari et al…”
Section: Experimental Assesmentmentioning
confidence: 99%
“…As Tiwari et al have noted, the losses incurred by evaluating several poor solutions can easily outweigh the benefits of discovering an excellent solution [12].…”
Section: Search Efficiencymentioning
confidence: 99%