Proceedings 20th IEEE International Parallel &Amp; Distributed Processing Symposium 2006
DOI: 10.1109/ipdps.2006.1639595
|View full text |Cite
|
Sign up to set email alerts
|

Conjugate gradient sparse solvers: performance-power characteristics

Abstract: We characterize the performance and power attributes of the conjugate gradient (CG) sparse solver which is widely used in scientific applications. We use cycle-accurate simulations with SimpleScalar and Wattch, on a processor and memory architecture similar to the configuration of a node of the BlueGene/L. We first demonstrate that substantial power savings can be obtained without performance degradation if low power modes of caches can be utilized. We next show that if Dynamic Voltage Scaling (DVS) can be use… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2007
2007
2014
2014

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(14 citation statements)
references
References 19 publications
0
13
0
Order By: Relevance
“…Then, using formula from (Jaleel et al, 2006, Hennessy andPatterson, 2003), the memory access time is represented as: is difficult in real computing environment. However, even with a simple memory prefetcher, the value is negligibly small in our algorithm since it accesses memory in sequential direction (Malkowski et al, 2005a;2005b).…”
Section: Multithreaded Iterative Solver: Mtcgmentioning
confidence: 99%
“…Then, using formula from (Jaleel et al, 2006, Hennessy andPatterson, 2003), the memory access time is represented as: is difficult in real computing environment. However, even with a simple memory prefetcher, the value is negligibly small in our algorithm since it accesses memory in sequential direction (Malkowski et al, 2005a;2005b).…”
Section: Multithreaded Iterative Solver: Mtcgmentioning
confidence: 99%
“…The matrix properties including the name, dimension (×10 3 ), number of nonzeros (×10 6 ) and the percentage of nonzeroes relative to a dense matrix of the same dimension are: bcsstk31, 35.6, 1.2, .09%; fdm2, 32.1,.16, .01%; qa8fm, 66.1, 1.6, .03%; and msc23052, 23.0, 1.1, .21%. These matrices had been reordered using the Reverse Cuthill McKee [17] scheme to improve the locality of access in the source vector [12] as is commonly done for tuned scientific codes.…”
Section: Sparse Scientific Computing Applicationsmentioning
confidence: 99%
“…Additionally, prefetching techniques were discussed by Lin, et al [31]. Effects of prefetchers on performance and power of sparse applications were investigated by authors in [12].…”
Section: Related Researchmentioning
confidence: 99%
See 1 more Smart Citation
“…We discuss how memory optimizations that we have developed earlier [11], [15], [16] can affect the performance of tuned and un-tuned versions of sparse matrix vector multiplication. We consider the use of such optimizations with powersaving modes of the hardware such as Dynamic Voltage and Frequency Scaling (DVFS) [5] to improve performance at significantly lower power levels.…”
Section: Introductionmentioning
confidence: 99%