2017
DOI: 10.1016/j.acha.2015.09.009
|View full text |Cite
|
Sign up to set email alerts
|

Multiscale geometric methods for data sets I: Multiscale SVD, noise and curvature

Abstract: Large data sets are often modeled as being noisy samples from probability distributions µ in R D , with D large. It has been noticed that oftentimes the support M of these probability distributions seems to be well-approximated by low-dimensional sets, perhaps even by manifolds. We shall consider sets that are locally well approximated by k-dimensional planes, with k ≪ D, with k-dimensional manifolds isometrically embedded in R D being a special case. Samples from µ are furthermore corrupted by D-dimensional n… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
53
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 53 publications
(53 citation statements)
references
References 68 publications
(127 reference statements)
0
53
0
Order By: Relevance
“…Our second goal is to obtain finite-sample results which hold outside the asymptotic regime. A common phenomenon in practice is for a measure to exhibit different dimensional structure at different scales; this so-called multi-scale behavior arises in a range of applications [LMR16,WDCB05,SMB98]. We show that the convergence ofμ n to µ in W p for such measures can exhibit wildly different rates as n increases.…”
Section: Introductionmentioning
confidence: 92%
“…Our second goal is to obtain finite-sample results which hold outside the asymptotic regime. A common phenomenon in practice is for a measure to exhibit different dimensional structure at different scales; this so-called multi-scale behavior arises in a range of applications [LMR16,WDCB05,SMB98]. We show that the convergence ofμ n to µ in W p for such measures can exhibit wildly different rates as n increases.…”
Section: Introductionmentioning
confidence: 92%
“…Principal component analysis (PCA) is the main representative of this class of algorithms. Both a global (gPCA) and a multiscale version (mPCA) of the algorithm are used [15,23]. In the former one evaluates the covariance matrix on the whole dataset X, whereas in the latter one performs the spectral analysis on local subsets X(x 0 , r c ) of X, obtained by selecting one particular point x 0 and including in the local covariance matrix only those points that lie inside a cutoff radius r c , which is then varied.…”
Section: Standard Algorithms For Ide and Extreme Locally Undersampledmentioning
confidence: 99%
“…However, the method presented above is global, thus it is expected to fail on manifolds with high intrinsic curvature. As in the case of the global PCA [15,23], we can overcome this issue by providing a suitable multiscale generalization of the FCI estimator. The multiscale FCI method selects single points in the dataset and their neighbours at a fixed maximum distance r cutoff , which is then varied.…”
Section: Robustness Of the Fci Estimatormentioning
confidence: 99%
“…The local covariance matrix and its relationship with the embedded tangent space of the manifold have been widely studied recently, including (but not exclusively) [23,5,28,3,12,14]. Recently, in order to systematically study the LLE algorithm, the higher order structure of the local covariance matrix was explored in [30].…”
Section: Local Covariance Matrix and Some Factsmentioning
confidence: 99%