2004
DOI: 10.1002/nla.382
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical hybrid grids: data structures and core algorithms for multigrid

Abstract: SUMMARYFor many scientiÿc and engineering applications, it is often desirable to use unstructured grids to represent complex geometries. Unfortunately, the data structures required to represent discretizations on such grids typically result in extremely ine cient performance on current high-performance architectures. Here, we introduce a grid framework using patch-wise, regular reÿnement that retains the exibility of unstructured grids, while achieving performance comparable to that seen with purely structured… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
84
0

Year Published

2007
2007
2021
2021

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 54 publications
(85 citation statements)
references
References 5 publications
1
84
0
Order By: Relevance
“…As the exchange and the optimization of the kernels for the various simulation scenarios and hardware is one features of our sweep concept, good performance results are obtained with the WaLBerla framework. In this respect, the framework benefits from our long-term experience in the optimization of numerical codes for serial [40][41][42][43] as well as parallel large scale simulations [44]. Performance results for WaLBerla can be found in [45].…”
Section: Efficiency and Scalabilitymentioning
confidence: 99%
“…As the exchange and the optimization of the kernels for the various simulation scenarios and hardware is one features of our sweep concept, good performance results are obtained with the WaLBerla framework. In this respect, the framework benefits from our long-term experience in the optimization of numerical codes for serial [40][41][42][43] as well as parallel large scale simulations [44]. Performance results for WaLBerla can be found in [45].…”
Section: Efficiency and Scalabilitymentioning
confidence: 99%
“…Serendipitously, this special processing order does not have a significant effect on the overall convergence of the solver; cf. [6]. For the new class of machines, additionally OpenMP parallelization was implemented to improve the node-level parallelism.…”
Section: Pressure-correction Schemementioning
confidence: 99%
“…The SR8000 is designed to be efficient even on codes that do not exhibit good spatial locality, but only when it can schedule the memory access using preload instructions. This makes the SR8000 extremely efficient on structured codes, where it is capable of achieving more than 50% of its Rpeak [2]. However, this architecture is also very sensitive to indirection, and, if it is not possible for the SR8000 to make efficient use of PVP, performance suffers.…”
Section: 11mentioning
confidence: 96%