2017
DOI: 10.1111/rssb.12226
|View full text |Cite
|
Sign up to set email alerts
|

Convex Clustering via l  1 Fusion Penalization

Abstract: We study the large sample behaviour of a convex clustering framework, which minimizes the sample within cluster sum of squares under an l 1 fusion constraint on the cluster centroids. This recently proposed approach has been gaining in popularity; however, its asymptotic properties have remained mostly unknown. Our analysis is based on a novel representation of the sample clustering procedure as a sequence of cluster splits determined by a sequence of maximization problems. We use this representation to provid… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
33
0
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 30 publications
(34 citation statements)
references
References 63 publications
0
33
0
1
Order By: Relevance
“…Our theoretical results are a significant extension of the univariate results in Radchenko and Mukherjee [43]. We rely herein on a more careful analysis of the empirical processes corresponding to the clustering trees constructed for each feature by the associated clustering procedure.…”
Section: Introductionmentioning
confidence: 67%
See 1 more Smart Citation
“…Our theoretical results are a significant extension of the univariate results in Radchenko and Mukherjee [43]. We rely herein on a more careful analysis of the empirical processes corresponding to the clustering trees constructed for each feature by the associated clustering procedure.…”
Section: Introductionmentioning
confidence: 67%
“…The convexity of the objective function in (1) has been exploited to develop algorithms for efficiently producing the path of solutions as a function of the penalty weight [31,32,43]. Clustering algorithms based on fusion penalization of this type have become very popular in large scale clustering [13,31,43,51,58] and regression analysis [8,37,47,48]. The entire path of solutions corresponding to the objective criterion (1) can be found by a simple merge algorithm in O(n ln n) operations.…”
Section: Methodology and Main Resultsmentioning
confidence: 99%
“…give sufficient conditions for exact cluster recovery at a fixed value of λ ("sparsistency"). Like us, Radchenko and Mukherjee (2017) are interested in properties of the entire solution path and give conditions under which convex clustering solutions (1) asymptotically yield the true dendrogram.…”
Section: Convex Clusteringmentioning
confidence: 99%
“…To address these limitations, several authors have recently studied a convex formulation of clustering (Pelckmans et al, 2005;. This convex formulation guarantees global optimality of the clustering solution and allows analysis of its theoretical properties (Tan and Witten, 2015;Radchenko and Mukherjee, 2017;Chi and Steinerberger, 2018). Despite these advantages, convex clustering has not yet achieved widespread popularity, due to its computationally intensive nature and lack of dendrogram-based visualizations.…”
Section: Introductionmentioning
confidence: 99%
“…These convex re-formulations have many noteworthy advantages over classical methods, including improved statistical performance, analytical tractability, and efficient and scalable computation. In particular, a convex formulation of clustering [1]- [3] has sparked much recent research, as it provides theoretical and computational guarantees not available for classical clustering methods [4]- [6]. Convex clustering combines a squared Frobenius norm loss term, which encourages the estimated centroids to remain near the original data, with a convex fusion penalty, typically the q -norm of the row-wise differences, which shrinks the estimated centroids together, inducing clustering behavior in the solution:…”
Section: Convex Clustering Bi-clustering and Co-clusteringmentioning
confidence: 99%