2017
DOI: 10.48550/arxiv.1708.03229
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Automatic Selection of t-SNE Perplexity

Yanshuai Cao,
Luyu Wang

Abstract: t-Distributed Stochastic Neighbor Embedding (t-SNE) is one of the most widely used dimensionality reduction methods for data visualization, but it has a perplexity hyperparameter that requires manual selection. In practice, proper tuning of t-SNE perplexity requires users to understand the inner working of the method as well as to have hands-on experience. We propose a model selection objective for t-SNE perplexity that requires negligible extra computation beyond that of the t-SNE itself. We empirically valid… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(17 citation statements)
references
References 3 publications
0
17
0
Order By: Relevance
“…For an illustration of cluster separation, we plot the t-distributed stochastic neighbor embedding (t-SNE) of a random sample 11 of 4400 devices in Figure 12 with perplexity chosen as in [21]. The clusters appear to be well separated with only minimal overlap, especially the single-company dominated cluster 1, thus reinforcing the clustering results.…”
Section: Temporal Traffic Spectrum and Clustering Analysismentioning
confidence: 56%
“…For an illustration of cluster separation, we plot the t-distributed stochastic neighbor embedding (t-SNE) of a random sample 11 of 4400 devices in Figure 12 with perplexity chosen as in [21]. The clusters appear to be well separated with only minimal overlap, especially the single-company dominated cluster 1, thus reinforcing the clustering results.…”
Section: Temporal Traffic Spectrum and Clustering Analysismentioning
confidence: 56%
“…• Scikit-learn package defaults for v0. 24.1, which are perplexity = 30, no exaggeration, and learning rate = 200 (note that we actually set it to 800 because we generate all embeddings using the OpenTSNE package, and in OpenTSNE the learning rate definition is 4 times smaller than in scikit-learn).…”
Section: Discussionmentioning
confidence: 99%
“…Because t-SNE has proven so popular, many researchers and software library authors have worked to identify guidelines for using t-SNE and selecting its hyperparameters [1,5,4,7,24,2,6]. Initially, researchers thought that t-SNE was robust to hyperparameter values, in particular for perplexity [1], but gradually research showed this is not completely true [4,6,5].…”
Section: Identifying Good T-sne Hyperparametersmentioning
confidence: 99%
See 2 more Smart Citations