2011
DOI: 10.1162/neco_a_00110
|View full text |Cite
|
Sign up to set email alerts
|

Divergence-Based Vector Quantization

Abstract: Supervised and unsupervised vector quantization methods for classification and clustering traditionally use dissimilarities, frequently taken as Euclidean distances. In this article, we investigate the applicability of divergences instead, focusing on online learning. We deduce the mathematical fundamentals for its utilization in gradient-based online vector quantization algorithms. It bears on the generalized derivatives of the divergences known as Fréchet derivatives in functional analysis, which reduces in … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
45
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
4
3
1

Relationship

2
6

Authors

Journals

citations
Cited by 65 publications
(45 citation statements)
references
References 40 publications
0
45
0
Order By: Relevance
“…In many applications, such as image analysis, pattern recognition and statistical machine learning we use the information-theoretic divergences rather than Euclidean squared or l p -norm distances [28]. Several information divergences such as Kullback-Leibler, Hellinger and Jensen-Shannon divergences are central to estimate similarity between distributions and have long history in information geometry.…”
Section: D(p || Z) ≤ D(p || Q) + D(q || Z) (Subaddivity/triangle Ineqmentioning
confidence: 99%
See 1 more Smart Citation
“…In many applications, such as image analysis, pattern recognition and statistical machine learning we use the information-theoretic divergences rather than Euclidean squared or l p -norm distances [28]. Several information divergences such as Kullback-Leibler, Hellinger and Jensen-Shannon divergences are central to estimate similarity between distributions and have long history in information geometry.…”
Section: D(p || Z) ≤ D(p || Q) + D(q || Z) (Subaddivity/triangle Ineqmentioning
confidence: 99%
“…Recently, alternative generalized divergences such as the Csiszár-Morimoto f -divergence and Bregman divergence become attractive alternatives for advanced machine learning algorithms [26][27][28][29][30][31][32][33][34]. In this paper, we discuss a robust parameterized subclass of the Csiszár-Morimoto and the Bregman divergences: Alpha-and Beta-divergences that may provide more robust solutions with respect to outliers and additive noise and improved accuracy.…”
Section: Introductionmentioning
confidence: 99%
“…Obviously, the Euclidean distance is not based on a functional norm [2,3,23]. Yet, the transfer to real functional norms and distances like Sobolev norms [24,25], the Lee-norm [23,1], kernel based LVQ-approaches [26] or divergence based similarity measures [27,28], which carry the functional aspect inherently, is straightforward and topic of future investigations.…”
Section: Resultsmentioning
confidence: 99%
“…Thereafter, g S ðtÞ is slowly increased in an adiabatic manner [17], such that all parameters can persistently follow the drift of the system. An additional term for b l -adaptation occurs for non-vanishing g S ðtÞ-values according to this new cost function (27):…”
Section: Structural Sparsitymentioning
confidence: 99%
“…It is interesting to consider other potential applications of Hölder divergences and compare their efficiency against the reference Cauchy-Schwarz divergence: For example, HD t-SNE (Stochastic Neighbor Embedding) compared to CS t-SNE [39], HD vector quantization (VQ) compared to CS VQ [40], HD saliency vs. CS saliency detection in images [41], etc.…”
mentioning
confidence: 99%