2015
DOI: 10.1093/bioinformatics/btv226
|View full text |Cite
|
Sign up to set email alerts
|

De novo meta-assembly of ultra-deep sequencing data

Abstract: We introduce a new divide and conquer approach to deal with the problem of de novo genome assembly in the presence of ultra-deep sequencing data (i.e. coverage of 1000x or higher). Our proposed meta-assembler Slicembler partitions the input data into optimal-sized ‘slices’ and uses a standard assembly tool (e.g. Velvet, SPAdes, IDBA_UD and Ray) to assemble each slice individually. Slicembler uses majority voting among the individual assemblies to identify long contigs that can be merged to the consensus assemb… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
12
0

Year Published

2015
2015
2021
2021

Publication Types

Select...
7
3

Relationship

1
9

Authors

Journals

citations
Cited by 25 publications
(13 citation statements)
references
References 26 publications
1
12
0
Order By: Relevance
“…However, consistent with our exploratory analysis here, some studies suggest that greater depth does not necessarily afford more contiguous assemblies and that ultra-deep sequencing (>1000× coverage) may actually be counterproductive if not explicitly handled using specialized assembly algorithms [171,172], due, in part, to the amplification of read duplication events and other sequencing errors [56]. From a practical standpoint, even if pre-processing steps successfully address the issues associated with ultra-deep sequencing, the significantly compounded cost of 10× more sequencing depth may not translate to greater information.…”
Section: Challenges In Biosynthetic Pathway Assembly and Product Psupporting
confidence: 73%
“…However, consistent with our exploratory analysis here, some studies suggest that greater depth does not necessarily afford more contiguous assemblies and that ultra-deep sequencing (>1000× coverage) may actually be counterproductive if not explicitly handled using specialized assembly algorithms [171,172], due, in part, to the amplification of read duplication events and other sequencing errors [56]. From a practical standpoint, even if pre-processing steps successfully address the issues associated with ultra-deep sequencing, the significantly compounded cost of 10× more sequencing depth may not translate to greater information.…”
Section: Challenges In Biosynthetic Pathway Assembly and Product Psupporting
confidence: 73%
“…It was released only shortly before SOAPdenovo2 was published and incorporates SOAPdenovo v1.05 and v1.06 as integral assembly components, using optimized parameters and revised error correction as well as scaffolding steps. SLICEMBLER [55] is a pipeline designed for ultra deep sequencing datasets using a “divide and conquer” approach. The read dataset is evenly divided into subsets, which are then assembled independently using an assembler of choice.…”
Section: Introductionmentioning
confidence: 99%
“…Also, for both SGVar and SGA-PDBG, there was a bigger loss of precision at 200×, compared with the lower coverage data ( Supplementary Figure S3 ). It has been found that the assembly quality would decrease once the sequencing depth goes too high [ 40–42 ], likely because of the accumulation of uncorrected sequencing errors.…”
Section: Resultsmentioning
confidence: 99%