2005
DOI: 10.1016/j.patcog.2004.03.024
|View full text |Cite
|
Sign up to set email alerts
|

Smoothing and compression with stochastic -testable tree languages

Abstract: In this paper, we describe some techniques to learn probabilistic k-testable tree models, a generalization of the well known k-gram models, that can be used to compress or classify structured data. These models are easy to infer from samples and allow for incremental updates. Moreover, as shown here, backing-off schemes can be defined to solve data sparseness, a problem that often arises when using trees to represent the data. These features make them suitable to compress structured data files at a better rate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2011
2011
2018
2018

Publication Types

Select...
2
2
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(9 citation statements)
references
References 36 publications
0
8
0
Order By: Relevance
“…The procedure to infer this kind of automata from a training sample, can be done easily (see [9] for details). This learning procedure can be extended to the case where the sample Ω is stochastically generated, incorporating probabilities to the DTA.…”
Section: Stochastic K-testable Tree Modelsmentioning
confidence: 99%
See 4 more Smart Citations
“…The procedure to infer this kind of automata from a training sample, can be done easily (see [9] for details). This learning procedure can be extended to the case where the sample Ω is stochastically generated, incorporating probabilities to the DTA.…”
Section: Stochastic K-testable Tree Modelsmentioning
confidence: 99%
“…As shown in [9], a probabilistic DTA (PDTA) incorporates a probability, p m (σ, t 1 , ..., t m ), for every transition in the automaton, with the normalization that the probabilities of the transitions leading to the same state q ∈ Q must add up to one. For this purpose, one should note that, in this kind of deterministic models, the likelihood of the training sample is maximized if the stochastic model assigns to every tree t in the sample a probability equal to its relative frequency in Ω [8].…”
Section: Stochastic K-testable Tree Modelsmentioning
confidence: 99%
See 3 more Smart Citations