2004
DOI: 10.1613/jair.1491
|View full text |Cite
|
Sign up to set email alerts
|

On Prediction Using Variable Order Markov Models

Abstract: This paper is concerned with algorithms for prediction of discrete sequences over a finite alphabet, using variable order Markov models. The class of such algorithms is large and in principle includes any lossless compression algorithm. We focus on six prominent prediction algorithms, including Context Tree Weighting (CTW), Prediction by Partial Match (PPM) and Probabilistic Suffix Trees (PSTs). We discuss the properties of these algorithms and compare their performance using real life sequences from three dom… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
286
0
1

Year Published

2007
2007
2015
2015

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 301 publications
(287 citation statements)
references
References 64 publications
0
286
0
1
Order By: Relevance
“…The crucial essence in compression is estimating the conditional probability for the next outcome given the past observations, so those symbols (and sub-sequences) with high conditional probabilities are assigned short codes . long sequences, attain the entropy lower bound (Begleiter et al, 2004). Thus, constructing a data-compression model that minimizes the average log-loss score of a sequence is equivalent to constructing a prediction model that maximizes the likelihood of a sequence.…”
Section: Introductionmentioning
confidence: 99%
See 4 more Smart Citations
“…The crucial essence in compression is estimating the conditional probability for the next outcome given the past observations, so those symbols (and sub-sequences) with high conditional probabilities are assigned short codes . long sequences, attain the entropy lower bound (Begleiter et al, 2004). Thus, constructing a data-compression model that minimizes the average log-loss score of a sequence is equivalent to constructing a prediction model that maximizes the likelihood of a sequence.…”
Section: Introductionmentioning
confidence: 99%
“…Therefore, in this study the terms "compression" and "prediction" are often considered as equivalent terms. Note that although other universal-compression algorithms can be used for prediction, the used VOM modela variation of Rissanen"s (1983) context tree -has been shown to attain the best asymptotic convergence rate for a given sequence (Ziv 2001, Begleiter et al, 2004.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations