2003
DOI: 10.1162/089976603321891819
|View full text |Cite
|
Sign up to set email alerts
|

Soft Learning Vector Quantization

Abstract: Learning vector quantization (LVQ) is a popular class of adaptive nearest prototype classifiers for multiclass classification, but learning algorithms from this family have so far been proposed on heuristic grounds. Here, we take a more principled approach and derive two variants of LVQ using a gaussian mixture ansatz. We propose an objective function based on a likelihood ratio and derive a learning rule using gradient descent. The new approach provides a way to extend the algorithms of the LVQ family to diff… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
129
0

Year Published

2006
2006
2016
2016

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 148 publications
(130 citation statements)
references
References 4 publications
1
129
0
Order By: Relevance
“…Generalized LVQ (GLVQ) [3] e.g., is based on a heuristic cost function which can be related to a maximization of the hypothesis margin of the classifier. Mathematically wellfounded alternatives were proposed in [4] and [14]: the cost functions of Soft LVQ and Robust Soft LVQ are based on a statistical modelling of the data distribution by a mixture of Gaussians, and training aims at optimizing the likelihood.…”
Section: Matrix Learning In Lvqmentioning
confidence: 99%
“…Generalized LVQ (GLVQ) [3] e.g., is based on a heuristic cost function which can be related to a maximization of the hypothesis margin of the classifier. Mathematically wellfounded alternatives were proposed in [4] and [14]: the cost functions of Soft LVQ and Robust Soft LVQ are based on a statistical modelling of the data distribution by a mixture of Gaussians, and training aims at optimizing the likelihood.…”
Section: Matrix Learning In Lvqmentioning
confidence: 99%
“…The RSLVQ algorithm [5] optimizes a log-likelihood loss under a framework of Gaussian mixture density:…”
Section: Nearest Prototype Classifiermentioning
confidence: 99%
“…High classification accuracies can be achieved by the NPC with a small number of prototypes when properly selected. The so-far proposed prototype learning methods can be grouped into ones based on heuristic adjustment (such as learning vector quantization (LVQ) [1]) and ones optimizing an objective function (such as [2,3,4,5,6]). These methods, optimizing a multi-class objective in training, do not consider outliers, which exist prevalently in applications of image recognition and document analysis, such as character string recognition [7], keyword retrieval [8] and transcript mapping [9].…”
Section: Introductionmentioning
confidence: 99%
“…If the width σ goes to zero, the assignment probabilities become hard assignments (winner-takes-all case) [5]. Smaller widths (σ → 0) and increasing distance to the cluster (d ij → ∞) are equivalent for this behavior of the Gaussian function: when σ approaches 0 the entire exponent d 2 ij /2σ 2 approaches ∞ just as the exponents in the equation above.…”
Section: B Influence Of the Similarity Functionmentioning
confidence: 99%
“…High numbers of irrelevant attributes should not cause clusters to get attracted or to coincide when f gauss is used. The abovementioned behavior of the Gaussian function can also be observed in soft learning vector quantization (SLVQ) when a Gaussian mixture approach is used [5]. In SLVQ the (un-normalized) assignment probabilities of a data point x j to prototype i are given by exp(d…”
Section: B Influence Of the Similarity Functionmentioning
confidence: 99%