1994
DOI: 10.1007/3-540-58473-0_141
|View full text |Cite
|
Sign up to set email alerts
|

Inducing probabilistic grammars by Bayesian model merging

Abstract: We describe a framework for inducing probabilistic grammars from corpora of positive samples. First, samples are incorporated by adding ad-hoc rules to a working grammar; subsequently, elements of the model (such as states or nonterminals) are merged to achieve generalization and a more compact representation. The choice of what to merge and when to stop is governed by the Bayesian posterior probability of the grammar given the data, which formalizes a trade-off between a close fit to the data and a default pr… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
105
0
1

Year Published

1997
1997
2016
2016

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 144 publications
(108 citation statements)
references
References 18 publications
(21 reference statements)
0
105
0
1
Order By: Relevance
“…For the HMM parameter estimation, we apply an incremental learning scheme utilizing the best first model merging framework [8,9]. Model merging is inspired by the observation that, when faced with new situations, humans and animals alike drive their learning process by first storing individual examples (memory based learning) when few data points are available and gradually switching to a parametric learning scheme to allow for better generalization as more and more data becomes available [10].…”
Section: Hmmmentioning
confidence: 99%
“…For the HMM parameter estimation, we apply an incremental learning scheme utilizing the best first model merging framework [8,9]. Model merging is inspired by the observation that, when faced with new situations, humans and animals alike drive their learning process by first storing individual examples (memory based learning) when few data points are available and gradually switching to a parametric learning scheme to allow for better generalization as more and more data becomes available [10].…”
Section: Hmmmentioning
confidence: 99%
“…Stolcke and Omohundro [58] propose a technique called Bayesian Model Merging (BMM): first, strings that are observed in the data are incorporated by adding ad-hoc rules to form an initial grammar; then, the grammar is made more concise by merging some of the rules. Stolcke and Omohundro [58] discuss two incarnations of their technique, one in which the models are probabilistic context-free grammars (PCFGs), and another in which they are hidden Markov models (HMMs).…”
Section: Computational Grammar Inductionmentioning
confidence: 99%
“…Stolcke and Omohundro [58] discuss two incarnations of their technique, one in which the models are probabilistic context-free grammars (PCFGs), and another in which they are hidden Markov models (HMMs). In the former, rules are merged by identifying non-terminal symbols A and B if the rule A → B is in the grammar; this leads to (over-) generalizations, and renders the grammar more compact.…”
Section: Computational Grammar Inductionmentioning
confidence: 99%
See 2 more Smart Citations