Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2013
DOI: 10.1145/2487575.2487591
|View full text |Cite
|
Sign up to set email alerts
|

Fast and scalable polynomial kernels via explicit feature maps

Abstract: Approximation of non-linear kernels using random feature mapping has been successfully employed in large-scale data analysis applications, accelerating the training of kernel machines. While previous random feature mappings run in O(ndD) time for n training samples in d-dimensional space and D random feature maps, we propose a novel randomized tensor product technique, called Tensor Sketching, for approximating any polynomial kernel in O(n(d + D log D)) time. Also, we introduce both absolute and relative error… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
242
1

Year Published

2014
2014
2022
2022

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 277 publications
(244 citation statements)
references
References 20 publications
1
242
1
Order By: Relevance
“…Many mappings have been proposed. Examples include random Fourier projection [83], random projections [84]- [87], polynomial approximation [88], and hashing [89]- [92]. They differ in various aspects, which are beyond the scope of this paper.…”
Section: B Approximation Of Kernel Methods Via Linear Classificationmentioning
confidence: 99%
“…Many mappings have been proposed. Examples include random Fourier projection [83], random projections [84]- [87], polynomial approximation [88], and hashing [89]- [92]. They differ in various aspects, which are beyond the scope of this paper.…”
Section: B Approximation Of Kernel Methods Via Linear Classificationmentioning
confidence: 99%
“…While the kernel trick has been widely and successfully applied in large margin learning, the calculation of kernel matrices is a bottleneck of the kernel trick for large-scale data sets. In recent years, a lot of alternatives to the kernel trick have been proposed to reduce the computational and storage costs (see, e.g., [43][44][45][46]), which can approximate the induced feature mapping φ by a low dimensional functionφ (x) such that…”
Section: Preliminariesmentioning
confidence: 99%
“…The recent fastfood algorithm [17] further speeds up RKS using matrix approximation techniques and reduces the time and space complexities. Other feature mapping techniques include those based on random projection [1,15,18,23], polynomial approximation [21], and hashing [19,32] Existing feature mapping techniques, when combined with linear classifiers, can achieve both nonlinear separability and higher scalability of linear classifiers. However, they cannot take advantage of the interpretability of linear classifiers.…”
Section: Related Workmentioning
confidence: 99%