2014
DOI: 10.1136/amiajnl-2013-002059
|View full text |Cite
|
Sign up to set email alerts
|

Leveraging the national cyberinfrastructure for biomedical research

Abstract: In the USA, the national cyberinfrastructure refers to a system of research supercomputer and other IT facilities and the high speed networks that connect them. These resources have been heavily leveraged by scientists in disciplines such as high energy physics, astronomy, and climatology, but until recently they have been little used by biomedical researchers. We suggest that many of the ‘Big Data’ challenges facing the medical informatics community can be efficiently handled using national-scale cyberinfrast… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
7
3

Relationship

1
9

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 8 publications
0
8
0
Order By: Relevance
“…Interestingly, exploiting big data technologies, such as large scale databases and parallel computing, is not only necessary for those online data repositories usually maintained by large research centers, but is also a need for any research lab that has NGS facilities. Building research management software nowadays is a big-data exercise, since these tools must manage large volumes of data of diverse nature, including raw signals, annotations, images, phenotypes, and textual reports [19,20,21]. Moreover, in particular for molecular diagnostic purposes, data are continuously generated and never erased, thus leading to a "Volume and Velocity" challenge to ensure accessibility of information.…”
Section: Big Data: Must-have or Nice-to-have?mentioning
confidence: 99%
“…Interestingly, exploiting big data technologies, such as large scale databases and parallel computing, is not only necessary for those online data repositories usually maintained by large research centers, but is also a need for any research lab that has NGS facilities. Building research management software nowadays is a big-data exercise, since these tools must manage large volumes of data of diverse nature, including raw signals, annotations, images, phenotypes, and textual reports [19,20,21]. Moreover, in particular for molecular diagnostic purposes, data are continuously generated and never erased, thus leading to a "Volume and Velocity" challenge to ensure accessibility of information.…”
Section: Big Data: Must-have or Nice-to-have?mentioning
confidence: 99%
“…For example, Precision Medicine Initiative (PMI) [ 2 ] aims to usher in a new era of medicine by collecting genomic data from a million people, by which more targeted treatment could be developed. It is becoming a big challenge to efficiently store and process the huge amount of genomic data in biomedical research [ 3 ]. Recently, cloud computing emerges [ 4 ] as an ideal platform for providing elastic computation and storage resources for genomic data analysis.…”
Section: Introductionmentioning
confidence: 99%
“…6 SRA contains 4.0 petabytes of data deposited in the last 6 years with geometric growth, 7 CGHub, 8 which stores National Cancer Institute data from petabyte scale projects including The Cancer Genomics Atlas 9 and Illumina's BaseSpace. 10 Data access at these repositories has inspired or taken advantage of advanced data transfer networks such as Internet2 100Gbit Advanced Layer 2 Services (AL2S 11 ) and GENI, 12 data flow control applications such as Globus Online 13 and Aspera, 14 and cyberinfrastructure environments (reviewed in 15 ) including XSEDE, 16 iPlant, 17 Open Science Grid, 18 ELIXIR, 19 UPP-NEX, 20 CloudBioLinux, 21 BGI-Cloud, 22 and CloudLab. 23 Diving sequencing costs and the power of the high-throughput sequencing measurement paradigm will surely accelerate genomics data accumulation and processing demands.…”
Section: The Widening Gulf Between Genomics Data Generation and Consumentioning
confidence: 99%