Proceedings of the Thirty-Seventh Annual ACM Symposium on Theory of Computing 2005
DOI: 10.1145/1060590.1060645
|View full text |Cite
|
Sign up to set email alerts
|

Learning nonsingular phylogenies and hidden Markov models

Abstract: In this paper we study the problem of learning phylogenies and hidden Markov models. We call a Markov model nonsingular if all transition matrices have determinants bounded away from 0 (and 1). We highlight the role of the nonsingularity condition for the learning problem. Learning hidden Markov models without the nonsingularity condition is at least as hard as learning parity with noise, a wellknown learning problem conjectured to be computationally hard. On the other hand, we give a polynomial-time algorithm… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

1
137
0

Year Published

2010
2010
2018
2018

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 114 publications
(138 citation statements)
references
References 34 publications
1
137
0
Order By: Relevance
“…A more ambitious problem is that of grammatical inference, where the goal is to induce the model only from sequences. Regarding spectral methods, Mossel and Roch (2005) study the induction of the topology of a phylogenetic tree-shaped model, and Hsu et al (2012) discuss spectral techniques to induce PCFG, with dependency grammars as a special case.…”
Section: Related Workmentioning
confidence: 99%
“…A more ambitious problem is that of grammatical inference, where the goal is to induce the model only from sequences. Regarding spectral methods, Mossel and Roch (2005) study the induction of the topology of a phylogenetic tree-shaped model, and Hsu et al (2012) discuss spectral techniques to induce PCFG, with dependency grammars as a special case.…”
Section: Related Workmentioning
confidence: 99%
“…A central goal of machine learning is to design efficient algorithms for fitting a model to a collection of observations. In recent years, there has been considerable progress on a variety of problems in this domain, including algorithms with provable guarantees for learning mixture models [1], [2], [3], [4], [5], phylogenetic trees [6], [7], HMMs [8], topic models [9], [10], and independent component analysis [11]. These algorithms crucially rely on the assumption that the observations were actually generated by a model in the family.…”
Section: A Backgroundmentioning
confidence: 99%
“…Both [14] and [8] stated as an open question the problem of obtaining a polynomial-time algorithm for learning a mixture of k > 2 product distributions. Indeed, recent work of Mossel and Roch [20] on learning phylogenetic trees argues that the rank-deficiency of transition matrices is a major source of difficulty, and this may indicate why k = 2 has historically been a barrier-a two-row matrix can be rank-deficient only if one row is a multiple of the other, whereas the general case of k > 2 is much more complex.…”
Section: Related Workmentioning
confidence: 99%
“…work centers on trying to find the exact best mixture model (in terms of likelihood) which explains a given data sample; this is computationally intractable in general. In contrast, our main goal (and the goal of [18,14,9,8,20]) is to obtain efficient algorithms that produce -close hypotheses.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation