2011
DOI: 10.1186/1471-2105-12-428
|View full text |Cite
|
Sign up to set email alerts
|

Fast MCMC sampling for hidden markov models to determine copy number variations

Abstract: BackgroundHidden Markov Models (HMM) are often used for analyzing Comparative Genomic Hybridization (CGH) data to identify chromosomal aberrations or copy number variations by segmenting observation sequences. For efficiency reasons the parameters of a HMM are often estimated with maximum likelihood and a segmentation is obtained with the Viterbi algorithm. This introduces considerable uncertainty in the segmentation, which can be avoided with Bayesian approaches integrating out parameters using Markov Chain M… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
24
1

Year Published

2012
2012
2020
2020

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 9 publications
(27 citation statements)
references
References 35 publications
2
24
1
Order By: Relevance
“…Similarly, HMM algorithms were originally formulated for aCGH platforms [22] and many innovations were subsequently proposed. For example, distance-based transition probabilities [6], fully Bayesian HMMs [23], reversible jump and approximate sampling Markov chain Monte Carlo (MCMC) [24,25], iterative approaches to parameter estimation [26], alternatives to the Viterbi algorithm [27], and higher order Markov chains [28]. As HMMs readily accomodate multiple data sequences, the observation that copy number can be estimated from genotyping arrays [29] led to the development of several HMMs that jointly model copy number and genotypes at SNPs [30-37].…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, HMM algorithms were originally formulated for aCGH platforms [22] and many innovations were subsequently proposed. For example, distance-based transition probabilities [6], fully Bayesian HMMs [23], reversible jump and approximate sampling Markov chain Monte Carlo (MCMC) [24,25], iterative approaches to parameter estimation [26], alternatives to the Viterbi algorithm [27], and higher order Markov chains [28]. As HMMs readily accomodate multiple data sequences, the observation that copy number can be estimated from genotyping arrays [29] led to the development of several HMMs that jointly model copy number and genotypes at SNPs [30-37].…”
Section: Introductionmentioning
confidence: 99%
“…Following [62], we report F-measures (F 1 scores) for binary classification into normal and aberrant segments ( Fig. 2), using the usual definition of F = 2πρ π+ρ being the harmonic mean of precision π = TP TP+FP and recall ρ = TP TP+FN , where TP, FP, TN and FN denote true/false positives/negatives, respectively.…”
Section: Simulated Acgh Datamentioning
confidence: 99%
“…Though there are several schemes available to sample q, [58] has argued strongly in favor of Forward-Backward sampling [57], which yields Forward-Backward Gibbs sampling (FBG) above. Variations of this have been implemented for segmentation of aCGH data before [60,62,78]. However, since in each iteration a quadratic number of terms has to be calculated at each position to obtain the forward variables, and a state has to be sampled at each position in the backward step, this method is still expensive for large data.…”
Section: Bayesian Hidden Markov Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Our approach uses a Poisson hidden Markov model (PHMM) to 1) estimate (hidden) states of gene expression levels in terminal exon 3′ UTRs, 2) infer shortening of the region in human liver and brain cortex tissues and 3) demonstrate tissue-specific APA. Others have used hidden Markov models (HMMs) in a similar fashion to dynamically map chromatin states (Ernst et al, 2011), to integrate genomic data (Day et al, 2007) and for determination of gene copy number variations (Mahmud and Schliep, 2011) just to name a few. We compare our results to those obtained by MISO, a probabilistic approach to quantification of transcripts at the 3′ UTR (Katz et al, 2010) and Cufflinks, based on de novo transcript assembly (Trapnell et al, 2010).…”
Section: Introductionmentioning
confidence: 99%