2013
DOI: 10.1007/978-3-642-40988-2_21
|View full text |Cite
|
Sign up to set email alerts
|

Inhomogeneous Parsimonious Markov Models

Abstract: Abstract. We introduce inhomogeneous parsimonious Markov models for modeling statistical patterns in discrete sequences. These models are based on parsimonious context trees, which are a generalization of context trees, and thus generalize variable order Markov models. We follow a Bayesian approach, consisting of structure and parameter learning. Structure learning is a challenging problem due to an overexponential number of possible tree structures, so we describe an exact and efficient dynamic programming al… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2014
2014
2020
2020

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 20 publications
0
7
0
Order By: Relevance
“…Even though not all TFs can be expected to bind to such a long binding site, adding some possibly uninformative positions is less harmful than not being able to take into account all informative positions. As a motif model we use inhomogeneous parsimonious Markov models [ 34 ] of order 0–4, which also includes the standard PWM model [ 1 , 2 ] that is equivalent to the PMM of order zero.…”
Section: Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…Even though not all TFs can be expected to bind to such a long binding site, adding some possibly uninformative positions is less harmful than not being able to take into account all informative positions. As a motif model we use inhomogeneous parsimonious Markov models [ 34 ] of order 0–4, which also includes the standard PWM model [ 1 , 2 ] that is equivalent to the PMM of order zero.…”
Section: Resultsmentioning
confidence: 99%
“…For estimating a sequence motif from limited training data, e.g., a few experimentally verified binding sites, the PWM model will certainly remain the optimal choice. For fully observable data, more than 10 2 training sequences are typically required before any model with more parameters than a PWM model can effectively utilize additional information in the data [ 34 , 42 ]. Moreover, due to its small parameter space, the PWM model is more robust to low-quality data, where models that attempt to infer dependencies may be more prone to adapting to noise.…”
Section: Discussionmentioning
confidence: 99%
See 3 more Smart Citations