2020
DOI: 10.5808/gi.2020.18.1.e8
|View full text |Cite
|
Sign up to set email alerts
|

Bioinformatics services for analyzing massive genomic datasets

Abstract: The explosive growth of next-generation sequencing data has resulted in ultra-large-scale datasets and ensuing computational problems. In Korea, the amount of genomic data has been increasing rapidly in the recent years. Leveraging these big data requires researchers to use large-scale computational resources and analysis pipelines. A promising solution for addressing this computational challenge is cloud computing, where CPUs, memory, storage, and programs are accessible in the form of virtual machines. Here,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 7 publications
(2 citation statements)
references
References 39 publications
0
2
0
Order By: Relevance
“…To tackle this predicament, researchers need solutions conducting genome wide calculations on expansive repositories in a fast, resource-efficient, and scalable nature (10,17,19). However, popular computational tools, though capable of being executed on HPC (High Performance Computing) clusters, have been designed with the goal of processing data that spans a few hundred individuals across thousands of polymorphic sites (20)(21)(22). They use single thread algorithms that do not exploit the prevalent parallel processing technologies made available through multi core CPUs (Control Processing Unit), and GPUs (Graphical Processing Unit) (10,21,23,24).…”
Section: Introductionmentioning
confidence: 99%
“…To tackle this predicament, researchers need solutions conducting genome wide calculations on expansive repositories in a fast, resource-efficient, and scalable nature (10,17,19). However, popular computational tools, though capable of being executed on HPC (High Performance Computing) clusters, have been designed with the goal of processing data that spans a few hundred individuals across thousands of polymorphic sites (20)(21)(22). They use single thread algorithms that do not exploit the prevalent parallel processing technologies made available through multi core CPUs (Control Processing Unit), and GPUs (Graphical Processing Unit) (10,21,23,24).…”
Section: Introductionmentioning
confidence: 99%
“…To tackle this predicament, researchers need solutions that are of a fast, resource‐efficient, and scalable nature (Pfeifer et al, 2014; Rozas et al, 2017; Szpiech & Hernandez, 2014). However, popular computational tools, though capable of being executed on HPC (high‐performance computing) clusters, have been designed with the goal of processing data that spans a few hundred individuals across thousands of polymorphic sites (Cook & Andersen, 2017; Ko et al, 2020; Siepel, 2019). They use single‐thread algorithms that do not exploit the prevalent parallel processing technologies made available through multi‐core CPUs (central processing unit), and GPUs (graphical processing unit) (Cook & Andersen, 2017; Ghorpade, 2012; Pfeifer et al, 2014; Tendler et al, 2002).…”
Section: Introductionmentioning
confidence: 99%