Proceedings DCC 2002. Data Compression Conference
DOI: 10.1109/dcc.2002.999978
|View full text |Cite
|
Sign up to set email alerts
|

A source coding approach to classification by vector quantization and the principle of minimum description length

Abstract: An algorithm for supervised classi cation using vector quantization and entropy coding is presented. The classi cation rule is formed from a set of training data fX i ; Y i g n i=1 , which are independent samples from a joint distribution P X Y . Based on the principle of Minimum Description Length MDL, a statistical model that approximates the distribution P X Y ought to enable ecient coding of X and Y . On the other hand, we expect a system that encodes X;Y e ciently to provide ample information on the distr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
1
0

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 17 publications
0
1
0
Order By: Relevance
“…For instance, when detecting a face in an image, features associated with the face often have a low-dimensional structure which is "embedded" as a submanifold in a cloud of essentially random features from the background. Model selection criteria such as the minimum description length (MDL) [28,22] serve as important modifications to MAP for estimating a model across classes of different complexity. MDL selects the model that minimizes the overall coding length of the given (training) data, hence the name "minimum description length" or "minimum coding length" [1].…”
Section: Issues With Learning the Distributions From Training Samplesmentioning
confidence: 99%
“…For instance, when detecting a face in an image, features associated with the face often have a low-dimensional structure which is "embedded" as a submanifold in a cloud of essentially random features from the background. Model selection criteria such as the minimum description length (MDL) [28,22] serve as important modifications to MAP for estimating a model across classes of different complexity. MDL selects the model that minimizes the overall coding length of the given (training) data, hence the name "minimum description length" or "minimum coding length" [1].…”
Section: Issues With Learning the Distributions From Training Samplesmentioning
confidence: 99%
“…However, it is hard to conclude they are with the same interestingness. Intuitively, c→ C 0 is the most interesting one, for it only contains 1 item in this antecedent and is favored by the Minimal Description Length Principle [8]. The situation for other four rules is more complicated and a systemic measure is deserved.…”
Section: Introductionmentioning
confidence: 99%