2009 IEEE International Symposium on Parallel &Amp; Distributed Processing 2009
DOI: 10.1109/ipdps.2009.5161074
|View full text |Cite
|
Sign up to set email alerts
|

Improving MPI-HMMER's scalability with parallel I/O

Abstract: As the size of biological sequence databases continues to grow, the time to search these databases has grown proportionally. This has led to many parallel implementations of common sequence analysis suites. However, it has become clear that many of these parallel sequence analysis tools do not scale well for medium to large-sized clusters. In this paper we describe an enhanced version of MPI-HMMER. We improve on MPI-HMMER's scalability through the use of parallel I/O and a parallel file system. Our enhancement… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2009
2009
2015
2015

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 12 publications
(16 reference statements)
0
11
0
Order By: Relevance
“…This increase in number of events led to the increase in overhead seen in Table 2. For example, the number of events sent by Gromacs [10] when it is solving a large problem using 128 cores is around two millions, which is more than twice as larger as the number of events sent by MPIHMMER [11] for solving a large problem. The larger number of events sent by Gromacs led to an almost 0.2% increase in the overhead imposed by PROMON .…”
Section: Monitoring Overheadmentioning
confidence: 99%
“…This increase in number of events led to the increase in overhead seen in Table 2. For example, the number of events sent by Gromacs [10] when it is solving a large problem using 128 cores is around two millions, which is more than twice as larger as the number of events sent by MPIHMMER [11] for solving a large problem. The larger number of events sent by Gromacs led to an almost 0.2% increase in the overhead imposed by PROMON .…”
Section: Monitoring Overheadmentioning
confidence: 99%
“…This “Disk Wall” problem will become worse along with the exponential growth of database sizes. Walters et al [ 12 ] studied the I/O problems in HMMER and enhanced its MPI implementation (called MPI-HMMER) using parallel I/O and a parallel file system. In this new version (called PIO-HMMER), the database distribution mechanism was modified so that the master node, instead of directly reading data and sending them to slave nodes, only distributes sequence indexes (each containing sequence offsets and lengths) to slave nodes and slave nodes read from the database in parallel.…”
Section: Software Accelerated Hmmermentioning
confidence: 99%
“…There has been a great deal of work on optimizing HMMER 2 for traditional parallel computers [13]- [16]. J. P. Walters and etc.…”
Section: Algorithm 1: Pseudo-code Of Hmmsearchmentioning
confidence: 99%