2019 IEEE Visualization Conference (VIS) 2019
DOI: 10.1109/visual.2019.8933568
|View full text |Cite
|
Sign up to set email alerts
|

Interpreting Distortions in Dimensionality Reduction by Superimposing Neighbourhood Graphs

Abstract: To perform visual data exploration, many dimensionality reduction methods have been developed. These tools allow data analysts to represent multidimensional data in a 2D or 3D space, while preserving as much relevant information as possible. Yet, they cannot preserve all structures simultaneously and they induce some unavoidable distortions. Hence, many criteria have been introduced to evaluate a map's overall quality, mostly based on the preservation of neighbourhoods. Such global indicators are currently use… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
2
2
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 9 publications
(12 citation statements)
references
References 16 publications
0
12
0
Order By: Relevance
“…The decomposition is usually done using a heatmap [61], Voronoi diagram [2,24,31], or 2D point-cloud [37,38]. By contrast, MING [10] explains False and Missing Neighbors by visualizing the shared amount of the nearest neighbor graphs constructed in the original and projected space. In this work, we quantified pointwise distortions by aggregating the inter-cluster distortion of the clusters and visualized them.…”
Section: Distortion Visualizationsmentioning
confidence: 99%
See 1 more Smart Citation
“…The decomposition is usually done using a heatmap [61], Voronoi diagram [2,24,31], or 2D point-cloud [37,38]. By contrast, MING [10] explains False and Missing Neighbors by visualizing the shared amount of the nearest neighbor graphs constructed in the original and projected space. In this work, we quantified pointwise distortions by aggregating the inter-cluster distortion of the clusters and visualized them.…”
Section: Distortion Visualizationsmentioning
confidence: 99%
“…For each pair of circles (A, B) centered at C A , C B , respectively, we adjusted the degree of overlap by changing C A OC B from 60°to 0°with an interval of 2.5°, where O is the origin. For each projection, we measured Steadiness and Cohesiveness (k = [80, 90, 100, 110, 120], 500 iterations), T&C (k = [5,10,15,20,15]), and MRREs (k = [5,10,15,20,15]). We used different k values and used the mean of their results as the final score for soundness.…”
Section: Experimental Designmentioning
confidence: 99%
“…Several efforts to characterize the quality of DR methods have been pursued [41,46,59], which can roughly be categorized as being global [30,53,[60][61][62][63][64][65][66][67][68] or local [29,45,46,69,70] in scope, and either based on preserving distances [68], neighborhoods [30, 46, 59-61, 63, 71, 72], or topology [67,73,74], but in all cases, they attempt to summarize the extent to which a given DR algorithm preserves some aspect of the original data's structure. In surveying this literature, and considering our basic principles, we find that what is still missing is an approach that not only assesses quality quantitatively and locally [45,47,50,[70][71][72][73]75], but also statistically in that it seeks to characterize the part of the natural and expected variability in quality that is due to noise.…”
Section: Assess Quality Statisticallymentioning
confidence: 99%
“…In traditional data analyses, statistics provides a rigorous framework with which to answer these questions, but DR methods confound the statistical distinction between signal and noise. Specifically, DR methods: generically produce distortions in their representations of data, and these distortions are inhomogeneous across a representation [30,40,[43][44][45][46][47]; are often stochastic and non-linear, meaning that the robustness and reproducibility of results is hard to assess [41]; and often require user specification of hyperparameters, where this specification is often based on heuristics rather than quantitative principles [10,[48][49][50]. Addressing these issues provides the motivation for this work, as recovering the ability to separate signal and noise in DR output is essential for their utilization in quantitative analyses.…”
Section: Introductionmentioning
confidence: 99%
“…In traditional data analyses, statistics provides a rigorous framework with which to answer these questions, but DR methods confound the statistical distinction between signal and noise. Specifically, DR methods generically produce distortions in their representations of data, and these distortions are inhomogeneous across a representation; 30 , 40 , 43 , 44 , 45 , 46 , 47 are often stochastic and non-linear, meaning that the robustness and reproducibility of results is hard to assess; 41 and often require user specification of hyperparameters, where this specification is often based on heuristics rather than quantitative principles. 10 , 48 , 49 , 50 Addressing these issues provides the motivation for this work, as recovering the ability to separate signal and noise in DR output is essential for their utilization in quantitative analyses.…”
Section: Introductionmentioning
confidence: 99%