2012 Data Compression Conference 2012
DOI: 10.1109/dcc.2012.40
|View full text |Cite
|
Sign up to set email alerts
|

Mixing Strategies in Data Compression

Abstract: We propose geometric weighting as a novel method to combine multiple models in data compression. Our results reveal the rationale behind PAQ-weighting and generalize it to a non-binary alphabet. Based on a similar technique we present a new, generic linear mixture technique. All novel mixture techniques rely on given weight vectors. We consider the problem of finding optimal weights and show that the weight optimization leads to a strictly convex (and thus, good-natured) optimization problem. Finally, an exper… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0
1

Year Published

2012
2012
2019
2019

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(15 citation statements)
references
References 7 publications
0
14
0
1
Order By: Relevance
“…Geometric mixing is an adaptive online ensemble that was analyzed in depth and whose properties are described in [24][25][26]. The main difference to linear mixing, which implies weighting the probabilities directly, is that the probabilities are first transformed into the logistic domain using the logit function (sometimes referred to as stretch in the paper).…”
Section: Context Mixingmentioning
confidence: 99%
“…Geometric mixing is an adaptive online ensemble that was analyzed in depth and whose properties are described in [24][25][26]. The main difference to linear mixing, which implies weighting the probabilities directly, is that the probabilities are first transformed into the logistic domain using the logit function (sometimes referred to as stretch in the paper).…”
Section: Context Mixingmentioning
confidence: 99%
“…In statistical data compression, one modeling approach used by many high performance programs is to use an ensemble method to combine the predictions of multiple statistical models (Mattern, 2012). Each model is typically tailored towards a particular kind of structure that occurs in popular file types.…”
Section: Introductionmentioning
confidence: 99%
“…Moreover, we add a theoretic justification and code length guarantees to PAQ's ad-hoc neural network mixing, since we show that it is a special form of the Geometric Mixture Distribution coupled with Online Gradient Descent. The results in Chapter 3 are a polished version of results published earlier in [58,59].…”
Section: Chapter 3: Elementary Modelingmentioning
confidence: 99%
“…Most of the results from this chapter have previously been published in [58,59]. In this chapter we present a greatly polished version thereof.…”
Section: Our Contributionmentioning
confidence: 99%
See 1 more Smart Citation