2023
DOI: 10.1093/biomet/asad004
|View full text |Cite
|
Sign up to set email alerts
|

A generalized Bayes framework for probabilistic clustering

Abstract: Loss-based clustering methods, such as k-means and its variants, are standard tools for finding groups in data. However, the lack of quantification of uncertainty in the estimated clusters is a disadvantage. Model-based clustering based on mixture models provides an alternative, but such methods face computational problems and large sensitivity to the choice of kernel. This article proposes a generalized Bayes framework that bridges between these paradigms through the use of Gibbs posteriors. In conducting Bay… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
15
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
(15 citation statements)
references
References 31 publications
0
15
0
Order By: Relevance
“…In those situations, even the definition of a “cluster” becomes less clear, so the choice of H$$ {H}_{\ell } $$ becomes problematic as well. We adhere to the philosophical viewpoint detailed in Rigon et al, 42 which is summarized in the following. We argue that it is often appropriate to treat the number of clusters as the resolution at which one wants to summarize the data rather than an unknown quantity one wishes to estimate.…”
Section: Discussionmentioning
confidence: 99%
“…In those situations, even the definition of a “cluster” becomes less clear, so the choice of H$$ {H}_{\ell } $$ becomes problematic as well. We adhere to the philosophical viewpoint detailed in Rigon et al, 42 which is summarized in the following. We argue that it is often appropriate to treat the number of clusters as the resolution at which one wants to summarize the data rather than an unknown quantity one wishes to estimate.…”
Section: Discussionmentioning
confidence: 99%
“…Partition-based clustering algorithms aim to minimize a specific loss function and are widely adopted but lack any quantification of uncertainty in the clustering solution. To address this and bridge the gap between partition-based and model-based approaches, the recent work of Rigon et al [ 79 ] employs a generalized Bayesian framework through the use of Gibbs posteriors [ 80 ]. Specifically, the generalized posterior is defined as where the loss function has the form with denoting the observations belonging to the th cluster and 0 quantifying the discrepancy of from the th cluster.…”
Section: Bayesian Cluster Analysismentioning
confidence: 99%
“…Specifically, the generalized posterior is defined as where the loss function has the form with denoting the observations belonging to the th cluster and 0 quantifying the discrepancy of from the th cluster. A simple example is the -means loss which sets (additional examples can be found in Rigon et al [ 79 ]). Fixing the number of clusters , the prior is chosen to be uniform over the set of partitions with clusters.…”
Section: Bayesian Cluster Analysismentioning
confidence: 99%
See 2 more Smart Citations