2011 IEEE International Parallel &Amp; Distributed Processing Symposium 2011
DOI: 10.1109/ipdps.2011.118
|View full text |Cite
|
Sign up to set email alerts
|

CATCH: A Cloud-Based Adaptive Data Transfer Service for HPC

Abstract: Abstract-Modern High Performance Computing (HPC)applications process very large amounts of data. A critical research challenge lies in transporting input data to the HPC center from a number of distributed sources, e.g., scientific experiments and web repositories, etc., and offloading the result data to geographically distributed, intermittently available endusers, often over under-provisioned connections. Such enduser data services are typically performed using point-to-point transfers that are designed for … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2011
2011
2017
2017

Publication Types

Select...
3
2
2

Relationship

2
5

Authors

Journals

citations
Cited by 22 publications
(10 citation statements)
references
References 12 publications
0
10
0
Order By: Relevance
“…ElasticSite [39] sent a part of the Grid workload to the Cloud when there is an overload of the user demand. CATCH [37] utilized the Cloud storage service for better data access between the desktop worker and the HPC center. Our work focuses on using the VC resources and only the missed deadline tasks are re-scheduled on the Cloud resources to improve the percentage of successful workflow.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…ElasticSite [39] sent a part of the Grid workload to the Cloud when there is an overload of the user demand. CATCH [37] utilized the Cloud storage service for better data access between the desktop worker and the HPC center. Our work focuses on using the VC resources and only the missed deadline tasks are re-scheduled on the Cloud resources to improve the percentage of successful workflow.…”
Section: Related Workmentioning
confidence: 99%
“…CyberShake [27] that is used by the Southern California Earthquake Center to characterize the earthquake hazards in a region. Sipht [28] which is used to automate the search for untranslated RNAs (sRNAs) for the bacterial replicons in the NCBI database, and LIGO [37] that is used to generate and analyze gravitational waveforms from data collected during the coalescing of compact binary systems. These scientific workflows are generated with 30 tasks by Bharathi et al [5,28].…”
Section: Workload Modelmentioning
confidence: 99%
“…We have also begun developing a decentralized data delivery service for HPC applications. For our initial investigation, we have relied on both cloud [13] and non-cloud [12] resources, and PlanetLab [1] for investigating the effectiveness of our approach.…”
Section: Preliminary Workmentioning
confidence: 99%
“…Approach feasibility using Azure: We have begun the integration of cloud and HPC by developing a data transfer service CATCH [13]. CATCH uses Azure [10] for cloud storage and FUSE [4] to provide HPC jobs a transparent file system mount point to access the cloud resources.…”
Section: Preliminary Workmentioning
confidence: 99%
“…In recent years, MapReduce/Hadoop [1] has emerged as the de facto model for big data applications, and is employed by industry [2], [3], [4], [5] and academia [6], [7], [8] alike. Improving the efficiency of Hadoop is therefore crucial.…”
Section: Introductionmentioning
confidence: 99%