2008 SC - International Conference for High Performance Computing, Networking, Storage and Analysis 2008
DOI: 10.1109/sc.2008.5219768
|View full text |Cite
|
Sign up to set email alerts
|

Toward loosely coupled programming on petascale systems

Abstract: Abstract-We have extended the Falkon execution framework to make loosely coupled petascale systems a practical and useful prog This work studies and measures the perf involved in applying this approach to enable th systems by a broader user community, and w Our work enables the execution of highly para composed of loosely coupled serial jobs with no the respective applications. This approach all potentially far larger-class of applications to l systems, such as the IBM Blue Gene/P sup present the challenges of… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
76
0

Year Published

2009
2009
2017
2017

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 75 publications
(76 citation statements)
references
References 19 publications
0
76
0
Order By: Relevance
“…Although no overhead is detected up to 256 CPUs, there is a maximum overhead of 7% (VEGFR2) at 4096 CPUs, which increases up to 16.3% (PARP) at 8192 CPUs. At 16,384 CPUs, the overhead varies from 27.8% (ACE) up to 34.3% (PARP). The performance of Autodock4.lga.MPI is similar to that of the program Dovis2 using 256 CPUs, 31 as well as the program Dock6.MPI that shows overheads of about 8% at 4096 CPUs, 12% at 8192 CPUs, and 45% at 16,384 CPUs.…”
Section: Scaling On Hpc Architecturementioning
confidence: 99%
See 3 more Smart Citations
“…Although no overhead is detected up to 256 CPUs, there is a maximum overhead of 7% (VEGFR2) at 4096 CPUs, which increases up to 16.3% (PARP) at 8192 CPUs. At 16,384 CPUs, the overhead varies from 27.8% (ACE) up to 34.3% (PARP). The performance of Autodock4.lga.MPI is similar to that of the program Dovis2 using 256 CPUs, 31 as well as the program Dock6.MPI that shows overheads of about 8% at 4096 CPUs, 12% at 8192 CPUs, and 45% at 16,384 CPUs.…”
Section: Scaling On Hpc Architecturementioning
confidence: 99%
“…14- 16 The performance can be significantly affected by time spent by the system in computations that do not contribute to the advancement of any user tasks (i.e., ''overhead''), such as task management, allocation of resources, and input/output (I/O) operations (open/read/write/close). [14][15][16] A task-parallel docking procedure (or high-throughput computing 17 ), i.e., n simultaneous and independent docking jobs running on n CPUs, can be carried out by executing a serial docking program (i.e., one that cannot run on more than one CPU) on each of the n CPUs.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…We view this approach as overly restrictive and potentially harmful in several ways: system reliability is jeopardized by more reboot cycles, diagnosing and monitoring the health of individual nodes is difficult, and the system is less available for use. Management based on virtualization would also make it possible to backfill work on the machine using loosely-coupled programming jobs [26] or other low priority work. A batch-submission or grid computing system could be run on a collection of nodes where a new OS stack could be dynamically launched; this system could also be brought up and torn down as needed.…”
Section: Motivationmentioning
confidence: 99%