2013
DOI: 10.1002/sam.11207
|View full text |Cite
|
Sign up to set email alerts
|

A distributed kernel summation framework for general‐dimension machine learning

Abstract: Kernel summations are a ubiquitous key computational bottleneck in many data analysis methods. In this paper, we attempt to marry, for the first time, the best relevant techniques in parallel computing, where kernel summations are in low dimensions, with the best general‐dimension algorithms from the machine learning literature. We provide the first distributed implementation of kernel summation framework that can utilize: (i) various types of deterministic and probabilistic approximations that may be suitable… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
10
0

Year Published

2015
2015
2023
2023

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 27 publications
0
10
0
Order By: Relevance
“…There are still interesting future directions to pursue, though. The first direction is parallelism: because our dual-tree algorithm is agnostic to the type of traversal used, we may use a parallel traversal (Curtin et al, 2013b), such as an adapted version of a recent parallel dual-tree algorithm (Lee et al, 2012). The second direction is kernel k-means and other spectral clustering techniques: our algorithm may be merged with the ideas of Curtin & Ram (2014) to perform kernel k-means.…”
Section: Discussionmentioning
confidence: 99%
“…There are still interesting future directions to pursue, though. The first direction is parallelism: because our dual-tree algorithm is agnostic to the type of traversal used, we may use a parallel traversal (Curtin et al, 2013b), such as an adapted version of a recent parallel dual-tree algorithm (Lee et al, 2012). The second direction is kernel k-means and other spectral clustering techniques: our algorithm may be merged with the ideas of Curtin & Ram (2014) to perform kernel k-means.…”
Section: Discussionmentioning
confidence: 99%
“…The dual-tree method is based on space partitioning trees for both the input sample and the evaluation points. These tree structures are then used to compute distances between input points and evaluation points more quickly, see Gray and Moore (2001), Gray and Moore (2003), Lang et al (2005), Lee et al (2006), Ram et al (2009), Curtin et al (2013), Griebel and Wissel (2013), Lee et al (2014). Among all these methods, the fast sum updating is the only one which is exact (no extra approximation is introduced) and whose speed is independent of the input data, the kernel and the bandwidth.…”
Section: Introductionmentioning
confidence: 99%
“…In the original FMM, kernel function is approximated by analytical tools (either with addition theorems of special functions or Taylor expansions) [12,4,6,10,5]. To overcome the difficulties when analytic formulation of kernel functions is not available, various semi-analytic [1,11,18] and algebraic FMMs [21,22,23] were developed in recent decades. In some other approaches [16,17], the whole kernel matrix is split into block matrices with various ranks, on each of which the SVD was implemented and then a truncated summation was used.…”
Section: Introductionmentioning
confidence: 99%