Proceedings of the Conference on High Performance Computing Networking, Storage and Analysis 2009
DOI: 10.1145/1654059.1654129
|View full text |Cite
|
Sign up to set email alerts
|

Scalable implicit finite element solver for massively parallel processing with demonstration to 160K cores

Abstract: Implicit methods for partial differential equations using unstructured meshes allow for an efficient solution strategy for many real-world problems (e.g., simulation-based virtual surgical planning). Scalable solvers employing these methods not only enable solution of extremely-large practical problems but also lead to dramatic compression in timeto-solution. We present a parallelization paradigm and associated procedures that enable our implicit, unstructured flow-solver to achieve strong scalability.We consi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
32
0

Year Published

2009
2009
2018
2018

Publication Types

Select...
4
4
1

Relationship

4
5

Authors

Journals

citations
Cited by 50 publications
(32 citation statements)
references
References 17 publications
0
32
0
Order By: Relevance
“…Since there is little computation performed during mesh adaptation relative to the substantial increase in communications required as the given mesh is distributed to more processors, the scaling decreases on high core counts (note that a strong scaling study is performed and therefore, the problem size is the same). However, the analysis have been shown to scale strongly with the similar amount of work load provided [21,41]. The fact that mesh modification routines are able to scale on bigger core count with more entities involved into communication supports the statement that it is likely to at a minimum provide the equivalent scaling with more work load on the same number of parts.…”
Section: Parallel Anisotropic Boundary Layer Adaptivity On a Heat Tramentioning
confidence: 62%
“…Since there is little computation performed during mesh adaptation relative to the substantial increase in communications required as the given mesh is distributed to more processors, the scaling decreases on high core counts (note that a strong scaling study is performed and therefore, the problem size is the same). However, the analysis have been shown to scale strongly with the similar amount of work load provided [21,41]. The fact that mesh modification routines are able to scale on bigger core count with more entities involved into communication supports the statement that it is likely to at a minimum provide the equivalent scaling with more work load on the same number of parts.…”
Section: Parallel Anisotropic Boundary Layer Adaptivity On a Heat Tramentioning
confidence: 62%
“…Both of these work types can be equally divided among the processors by partitioning the aggregate mesh into equal load parts [12,13]. So far, PHASTA is a pure MPI based code and each process executes of copy of the analysis code to handle the computation work and interactions corresponding to its mesh part.…”
Section: Parallelizationmentioning
confidence: 99%
“…The more complex message passing parallel model employs peer-to-peer (p2p) message exchanges among processors and takes advantage of communication and computation overlapping. Following [4], the point-to-point (p2p) communication strategy is based on a master-slave relationship between processors. This relationship is established by creating a hierarchy based on host partition numbers.…”
Section: Edgecfd: the Benchmark Softwarementioning
confidence: 99%