Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence 2017
DOI: 10.24963/ijcai.2017/207
|View full text |Cite
|
Sign up to set email alerts
|

Data-driven Random Fourier Features using Stein Effect

Abstract: Large-scale kernel approximation is an important problem in machine learning research. Approaches using random Fourier features have become increasingly popular, where kernel approximation is treated as empirical mean estimation via Monte Carlo (MC) or Quasi-Monte Carlo (QMC) integration. A limitation of the current approaches is that all the features receive an equal weight summing to 1. In this paper, we propose a novel shrinkage estimator from ''Stein effect", which provides a data-driven weighting strategy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 15 publications
(6 citation statements)
references
References 1 publication
0
6
0
Order By: Relevance
“…weighted random features: [33], [82] for RFF, [26] for QMC, [27] for GQ kernel alignment: KA-RFF [83] and KP-RFF [45] compressed low-rank approximation: CLR-RFF [47] kernel learning by random features…”
Section: Taxonomy Of Random Features Based Algorithmsmentioning
confidence: 99%
See 1 more Smart Citation
“…weighted random features: [33], [82] for RFF, [26] for QMC, [27] for GQ kernel alignment: KA-RFF [83] and KP-RFF [45] compressed low-rank approximation: CLR-RFF [47] kernel learning by random features…”
Section: Taxonomy Of Random Features Based Algorithmsmentioning
confidence: 99%
“…ii) Re-weighted random feature selection: Here the basic idea is to re-weight the random features by solving a constrained optimization problem. Examples of this approach include weighted RFF [33], [82], weighted QMC [26], and weighted GQ [27]. Note that these algorithms directly learn the weights of pre-given random features.…”
Section: Taxonomy Of Random Features Based Algorithmsmentioning
confidence: 99%
“…Similarly, sampling from an importance-weighted distribution may also be used in low-rank matrix approximation, i.e., column sampling, but algorithms in the setting of the column sampling [45][46][47] are not applicable to RFs [9]. Quasi-Monte Carlo techniques [48,49] can also improve M , but it is unknown whether they can achieve minimal M . In contrast, our algorithm achieves minimal M within feasible runtime.…”
Section: Generalization Property and Runtime Of Classification With O...mentioning
confidence: 99%
“…Another line of research has focused on data-dependent choice of random features. In [30,31,32,33], data-dependent random features has been studied for the approximation of shiftinvariant/translation-invariant kernels. On the other hand, in [34,35,36,37], the focal point is on the improvement of the out-of-sample error.…”
Section: Related Literaturementioning
confidence: 99%