2012
DOI: 10.1109/tvcg.2011.24
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid Parallelism for Volume Rendering on Large-, Multi-, and Many-Core Systems

Abstract: Abstract-With the computing industry trending towards multi-and many-core processors, we study how a standard visualization algorithm, ray-casting volume rendering, can benefit from a hybrid parallelism approach. Hybrid parallelism provides the best of both worlds: using distributed-memory parallelism across a large numbers of nodes increases available FLOPs and memory, while exploiting shared-memory parallelism among the cores within each node ensures that each node performs its portion of the larger calculat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
44
0

Year Published

2012
2012
2016
2016

Publication Types

Select...
8
2

Relationship

0
10

Authors

Journals

citations
Cited by 56 publications
(44 citation statements)
references
References 19 publications
(25 reference statements)
0
44
0
Order By: Relevance
“…Although it is reasonable to ignore the shared memory between the four cores on Intrepid, future computers will have many more cores per node. Some introductory work has analyzed the behavior of image compositing in shared-memory architectures [7,19,21,23], but further refinement is required to take advantage of the hybrid distributed memory plus shared memory architecture of large systems and to evolve the compositing as architectures and rendering algorithms change.…”
Section: Discussionmentioning
confidence: 99%
“…Although it is reasonable to ignore the shared memory between the four cores on Intrepid, future computers will have many more cores per node. Some introductory work has analyzed the behavior of image compositing in shared-memory architectures [7,19,21,23], but further refinement is required to take advantage of the hybrid distributed memory plus shared memory architecture of large systems and to evolve the compositing as architectures and rendering algorithms change.…”
Section: Discussionmentioning
confidence: 99%
“…Our results suggest a wide variation in performance can result -up to 254% on multi-core CPUs and 265% on many-core GPUs. We used the findings of this study to set tunable algorithmic parameters for a set of extreme-concurrency runs [18,19] that required literally millions of CPU hours; by finding and using optimal settings for tunable algorithmic parameters, we in effect saved millions of additional CPU hours that would have been spent executing an application in a non-optimal configuration. This work, which uses a well-established methodology for finding optimal performance, shows that such a methodology can be useful for visualization algorithms as well, and that the algorithmic parameters that produce the best performance vary from problem to problem and platform to platform, often in a non-obvious way.…”
Section: Discussionmentioning
confidence: 99%
“…Another class of large-scale volume rendering systems is purely CPU-based, in order to avoid GPU memory limitations altogether. Much research has been devoted to volume rendering on large supercomputers [5,16,26]. This is especially useful in the context of in-situ visualization of large-scale simulations, where the visualization is computed on the same machine as the data, avoiding the need to move large data.…”
Section: Related Workmentioning
confidence: 99%