1984
DOI: 10.1109/tcom.1984.1096090
|View full text |Cite
|
Sign up to set email alerts
|

Data Compression Using Adaptive Coding and Partial String Matching

Abstract: The recently developed technique of arithmetic coding, in conjunction with a Markov model of the source, is a powerful method of data compression in situations where a linear treatment is inappropriate. Adaptive coding allows the model to be constructed dynamically by both encoder and decoder during the course of the transmission, and has been shown to incur a smaller coding overhead than explicit transmission of the model's statistics. But there is a basic conflict between the desire to use high-order Markov … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
575
0
6

Year Published

1992
1992
2016
2016

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 977 publications
(584 citation statements)
references
References 14 publications
3
575
0
6
Order By: Relevance
“…In [14] text compression methods are considered for its extension to bitext compression considering exact correspondences between two words, and synonymy relationships between the words in both texts (as given by a thesaurus). These parallel predictions are then combined with PPM [3] ones. The weighting of both models are carefully tuned improving PPM compression ratios on separate texts.…”
Section: Compression Of Bitextsmentioning
confidence: 99%
See 1 more Smart Citation
“…In [14] text compression methods are considered for its extension to bitext compression considering exact correspondences between two words, and synonymy relationships between the words in both texts (as given by a thesaurus). These parallel predictions are then combined with PPM [3] ones. The weighting of both models are carefully tuned improving PPM compression ratios on separate texts.…”
Section: Compression Of Bitextsmentioning
confidence: 99%
“…-a Spanish-Catalan (es-ca) bitext from El Periódico de Catalunya, 2 a daily newspaper published both in Catalan and Spanish; -a Spanish-Galician (es-gl) bitext from Diario Oficial de Galicia, 3 the bulletin of the Government of Galicia, published both in Galician and Spanish; and -bitexts for German-English (de-en), Spanish-English (es-en) and French-English (fr-en) from the European Parliament Proceedings Parallel Corpus [8].…”
Section: Searching All the Possible Translations Of A Wordmentioning
confidence: 99%
“…Each user took dictation from Jane Austen's Emma in five-minute sessions. The language model (PPMD5) predicts the next character given the previous five characters; 6,7 it was trained on passages from Emma not included in the dictation. Right panels, the two experts took dictation using the same eyetracker to control the WiViK on-screen keyboard (a standard qwerty keyboard) with the word-completion buttons enabled.…”
Section: Commentioning
confidence: 99%
“…The leading data representation method for compression purposes is Huffman coding, which forms the basis of most subsequent approaches [20]. Compression techniques can be broadly classified in four major categories: derivatives of Lempel-Ziv-Welch [23], approaches based on statistical model prediction [4], on characters permutations [1], and on arithmetic coding [16]. In the case of DEM data, the seminal work [12] advocated the possibility of reducing data size with an initial data simplification stage, followed by compression with Huffman coding.…”
Section: Related Workmentioning
confidence: 99%
“…Prediction by Partial Matching (PPM) [4] algorithms consider the correlation between values (which, for example, could be linearly growing). They use N past values to predict the next one, trying to find the best relationship.…”
Section: Currently There Are Two Classes Of Algorithms That Perform mentioning
confidence: 99%