Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence 2021
DOI: 10.24963/ijcai.2021/647
|View full text |Cite
|
Sign up to set email alerts
|

Improved Guarantees and a Multiple-descent Curve for Column Subset Selection and the Nystrom Method (Extended Abstract)

Abstract: The Column Subset Selection Problem (CSSP) and the Nystrom method are among the leading tools for constructing interpretable low-rank approximations of large datasets by selecting a small but representative set of features or instances. A fundamental question in this area is: what is the cost of this interpretability, i.e., how well can a data subset of size k compete with the best rank k approximation? We develop techniques which exploit spectral properties of the data matrix to obtain improved approximation … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(21 citation statements)
references
References 32 publications
0
11
0
Order By: Relevance
“…Similarly, the recent advances in using and understanding the Nyström method (Williams and Seeger, 2001), which is one of the most popular sparse approximations in kernel methods, have been made independently to those of sparse GP approximations. The majority of these advances focus on an efficient approximation of the kernel matrix (e.g., Drineas and Mahoney, 2005;Belabbas and Wolfe, 2009;Gittens and Mahoney, 2016;Derezinski et al, 2020) or empirical risk minimization in the RKHS with a reduced basis (e.g, Bach, 2013;El Alaoui and Mahoney, 2015;Rudi et al, 2015Rudi et al, , 2017Meanti et al, 2020). This separation of two lines of research are arguably due to the difference in the notations and modeling philosophies of GPs and kernel methods.…”
Section: Introductionmentioning
confidence: 99%
“…Similarly, the recent advances in using and understanding the Nyström method (Williams and Seeger, 2001), which is one of the most popular sparse approximations in kernel methods, have been made independently to those of sparse GP approximations. The majority of these advances focus on an efficient approximation of the kernel matrix (e.g., Drineas and Mahoney, 2005;Belabbas and Wolfe, 2009;Gittens and Mahoney, 2016;Derezinski et al, 2020) or empirical risk minimization in the RKHS with a reduced basis (e.g, Bach, 2013;El Alaoui and Mahoney, 2015;Rudi et al, 2015Rudi et al, , 2017Meanti et al, 2020). This separation of two lines of research are arguably due to the difference in the notations and modeling philosophies of GPs and kernel methods.…”
Section: Introductionmentioning
confidence: 99%
“…The initial Nyström samples  (0) we considered were draw uniformly at random without replacement; while our experiments suggest that the local minima of the radial SKD often induce approximations of comparable quality, the use of more efficient initialisation strategies may be investigated (see e.g. [3,4,11,13,18]).…”
Section: Discussionmentioning
confidence: 99%
“…For a Nyström sample  ∈ of size ∈ ℕ, the matrix ̂ () is of rank at most . Following [4,10], to further assess the efficiency of the approximation of induced by , we introduce the approximation factors…”
Section: Numerical Experimentsmentioning
confidence: 99%
See 1 more Smart Citation
“…Kumar et al (2012) explored the sampling approach for the column subset selection problem by the Nyström method. Derezinski et al (2020) recently provided an improved theoretical guarantee for low-rank approximations of large datasets. Another popular idea in machine learning is coreset, which constructs estimators based on sub-data.…”
Section: Introductionmentioning
confidence: 99%