2010 ACM/IEEE International Conference for High Performance Computing, Networking, Storage and Analysis 2010
DOI: 10.1109/sc.2010.28
|View full text |Cite
|
Sign up to set email alerts
|

Functional Partitioning to Optimize End-to-End Performance on Many-core Architectures

Abstract: Abstract-Scaling computations on emerging massive-core supercomputers is a daunting task, which coupled with the significantly lagging system I/O capabilities exacerbates applications' end-to-end performance. The I/O bottleneck often negates potential performance benefits of assigning additional compute cores to an application. In this paper, we address this issue via a novel functional partitioning (FP) runtime environment that allocates cores to specific application tasks -checkpointing, de-duplication, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
5
4
1

Relationship

3
7

Authors

Journals

citations
Cited by 57 publications
(30 citation statements)
references
References 35 publications
(39 reference statements)
0
30
0
Order By: Relevance
“…The use of dedicated I/O cores [5], [26]- [28], threads [29], [30], or dedicated nodes [31], [32] (also termed "staging areas") is becoming more and more common. These strategies overlap I/O with computation by shipping data to dedicated resources, and offer more liberty in delaying actual I/O accesses.…”
Section: B Application-side I/o Schedulingmentioning
confidence: 99%
“…The use of dedicated I/O cores [5], [26]- [28], threads [29], [30], or dedicated nodes [31], [32] (also termed "staging areas") is becoming more and more common. These strategies overlap I/O with computation by shipping data to dedicated resources, and offer more liberty in delaying actual I/O accesses.…”
Section: B Application-side I/o Schedulingmentioning
confidence: 99%
“…Space-partitioning using dedicated cores to handle I/O or visualization tasks has been proposed using a FUSE interface [16] or an active buffering scheme for collective I/O [20]. The use of a FUSE interface produces multiple copies of data passing through the kernel space, increasing memory usage.…”
Section: Tightly-coupled Isv: Challenges and Solutionsmentioning
confidence: 99%
“…In our design, the client and center stubs talk to a transparent file system mount point, provided through FUSE [8] as Cloud FS (Figure 1), which abstracts the process of accessing the cloud storage and in addition moves the data closer to the end-user or the HPC center. The use of FUSE to abstract access to different storage substrates has gained wide spread popularity due to the ease with which purpose-built storage systems can be transparently made available by having them implement certain POSIX APIs (e.g., s3fs [15] for Amazon S3 or stdchk [16], [17], a file system atop distributed storage of disks, memory or SSD.) The read() or write() call in these situations typically abstracts parallel striping or a network transfer, respectively.…”
Section: Data Transport As a File Systemmentioning
confidence: 99%