2009 IEEE International Symposium on Parallel &Amp; Distributed Processing 2009
DOI: 10.1109/ipdps.2009.5161070
|View full text |Cite
|
Sign up to set email alerts
|

Making resonance a common case: A high-performance implementation of collective I/O on parallel file systems

Abstract: Collective I/O is a widely used technique to improve I/O performance in parallel computing. It can be implemented as a client-based or server-based scheme. The client-based implementation is more widely adopted in MPI-IO software such as ROMIO because of its independence from the storage system configuration and its greater portability. However, existing implementations of client-side collective I/O do not take into account the actual pattern of file striping over multiple I/O nodes in the storage system. This… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
21
0

Year Published

2010
2010
2024
2024

Publication Types

Select...
4
2
1

Relationship

3
4

Authors

Journals

citations
Cited by 40 publications
(21 citation statements)
references
References 14 publications
(21 reference statements)
0
21
0
Order By: Relevance
“…However, if the requests to a disk are issued from different aggregators, there is no way to ensure ascending order unless the involved aggregators synchronize their issuance, which can be excessively expensive, especially in a large-scale system. The ideal scenario in terms of maximizing disk efficiency is to have all requests to a disk (or data node) be issued by one aggregator that can send them in the ascending order according to requested data's offsets in the file [35].…”
Section: Potential Challenges In the Performance Of Collective I/omentioning
confidence: 99%
See 2 more Smart Citations
“…However, if the requests to a disk are issued from different aggregators, there is no way to ensure ascending order unless the involved aggregators synchronize their issuance, which can be excessively expensive, especially in a large-scale system. The ideal scenario in terms of maximizing disk efficiency is to have all requests to a disk (or data node) be issued by one aggregator that can send them in the ascending order according to requested data's offsets in the file [35].…”
Section: Potential Challenges In the Performance Of Collective I/omentioning
confidence: 99%
“…While maintaining an equal number of aggregators as data nodes allows the data nodes to receive fully sorted sequences of requests, it may limit the I/O bandwidth if the number of data nodes is small. One solution is having multiple aggregators coordinated to access a data node [35].…”
Section: Candidate Collective-i/o Implementationsmentioning
confidence: 99%
See 1 more Smart Citation
“…While collective I/O can incur communication overhead because of data exchange among processes, its performance advantage is well recognized, making it one of most popular I/O optimization techniques for MPI programs. Observing that current ROMIO implementations of collective I/O [28] can cause requests to arrive at each data server in an order inconsistent with data placement, resonant I/O was proposed as an enhanced implementation of collective I/O to restore spatial locality [34].…”
Section: A Modifying Request Streams For Greater Spatial Localitymentioning
confidence: 99%
“…How to optimize I/O performance is elusive, and the optimization is a complex, error-prone, and time-consuming task, especially for applications with complex I/O behaviors. For example, Zhang's work [7] shows that Collective I/O works well in some cases but not in others. Song's work [8] shows that finding the optimal data layout configuration in PVFS2 can be a daunting task.…”
Section: Introductionmentioning
confidence: 99%