2014
DOI: 10.3390/e16063273
|View full text |Cite
|
Sign up to set email alerts
|

On Clustering Histograms with k-Means by Using Mixed α-Divergences

Abstract: Clustering sets of histograms has become popular thanks to the success of the generic method of bag-of-X used in text categorization and in visual categorization applications. In this paper, we investigate the use of a parametric family of distortion measures, called the α-divergences, for clustering histograms. Since it usually makes sense to deal with symmetric divergences in information retrieval systems, we symmetrize the α-divergences using the concept of mixed divergences. First, we present a novel exten… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 29 publications
(20 citation statements)
references
References 31 publications
0
19
0
Order By: Relevance
“…Another interesting case of study is the -divergence that can be obtained from the -divergences for the parametrization . As mentioned before, there are closed formulas for the computation of sided centroids for the k -means with -divergences [ 7 , 9 ]. However, to compare the formulas, it is necessary to take into account that there are two equivalent ways to define the -divergence family.…”
Section: Relations With Known Centroid-based Algorithmsmentioning
confidence: 99%
See 1 more Smart Citation
“…Another interesting case of study is the -divergence that can be obtained from the -divergences for the parametrization . As mentioned before, there are closed formulas for the computation of sided centroids for the k -means with -divergences [ 7 , 9 ]. However, to compare the formulas, it is necessary to take into account that there are two equivalent ways to define the -divergence family.…”
Section: Relations With Known Centroid-based Algorithmsmentioning
confidence: 99%
“…For the specific case of -divergences, closed formulas for the computation of the sided centroids were derived in [ 6 ] for the right-type, and in [ 7 , 8 ] for the left-type. Symmetrized centroids have also been derived for clustering histograms in the Bag-of-Words modeling paradigm in [ 9 ]. Total Bregman divergences (TDB), which are invariant to particular transformations on the natural space, have also been used for estimating the center of a set of vectors in [ 10 ] in the context of the shape retrieval problem.…”
Section: Introductionmentioning
confidence: 99%
“…We use JS-divergence as the objective function dq in our experiments (see Section 5). JS-divergence is a standard measure for quantifying distances between probability distributions, which is often used in histogram/vector classification [45] and clustering [51]. Given two histograms H 1 , H 2 , the JS-divergence between them is defined as…”
Section: Quality Lossmentioning
confidence: 99%
“…The k-means clustering algorithm [ 9 , 10 , 11 ] is a hard clustering algorithm, which is representative of a typical prototype-based clustering method for objective functions. It utilizes the distance from data points to the prototype as an optimized objective function and an adjustment rule of iterative operation is obtained by using the method of function extreme value evaluation [ 12 , 13 , 14 , 15 ]. Entropy is a method of measuring and quantitatively describing the randomness or irregularity of a time series or nonlinear signals [ 16 ].…”
Section: Introductionmentioning
confidence: 99%