2008
DOI: 10.1007/978-3-540-87481-2_10
|View full text |Cite
|
Sign up to set email alerts
|

Support Vector Machines, Data Reduction, and Approximate Kernel Matrices

Abstract: Abstract. The computational and/or communication constraints associated with processing large-scale data sets using support vector machines (SVM) in contexts such as distributed networking systems are often prohibitively high, resulting in practitioners of SVM learning algorithms having to apply the algorithm on approximate versions of the kernel matrix induced by a certain degree of data reduction. In this paper, we study the tradeoffs between data reduction and the loss in an algorithm's classification perfo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

2011
2011
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 17 publications
(8 citation statements)
references
References 13 publications
(15 reference statements)
0
8
0
Order By: Relevance
“…In fact a hyperplane only requires the determination of two parameters, that is, a weight vector (that determines its slope) and a bias parameter. Depending on the number of target classes, SVM based techniques can be classified as (Tax and Duin 1999;Aly 2005;Bishop 2006;Nguyen et al 2008;Yeung et al 2007;Steinwart and Christmann 2008;Tax and Duin 2004;Scholkopf and Smola 2001;Herbrich 2002;Weston and Watkins 1998;Smola and Schölkopf 1998;Weston and Watkins 1999;Mayoraz and Alpaydin 1999;Bredensteiner and Bennett 1999;Schwenker 2000;Schölkopf et al 2001;Hsu and Lin 2002;Franc and Hlavác 2002;Elisseeff and Weston 2002;Zhu et al 2003 Outlier and event detection in WSNs require a model of normal data. All data samples which do not fit the normal data model are declared to be outliers.…”
Section: Bayesian Based Approachesmentioning
confidence: 98%
See 1 more Smart Citation
“…In fact a hyperplane only requires the determination of two parameters, that is, a weight vector (that determines its slope) and a bias parameter. Depending on the number of target classes, SVM based techniques can be classified as (Tax and Duin 1999;Aly 2005;Bishop 2006;Nguyen et al 2008;Yeung et al 2007;Steinwart and Christmann 2008;Tax and Duin 2004;Scholkopf and Smola 2001;Herbrich 2002;Weston and Watkins 1998;Smola and Schölkopf 1998;Weston and Watkins 1999;Mayoraz and Alpaydin 1999;Bredensteiner and Bennett 1999;Schwenker 2000;Schölkopf et al 2001;Hsu and Lin 2002;Franc and Hlavác 2002;Elisseeff and Weston 2002;Zhu et al 2003 Outlier and event detection in WSNs require a model of normal data. All data samples which do not fit the normal data model are declared to be outliers.…”
Section: Bayesian Based Approachesmentioning
confidence: 98%
“…Labeled input-output pairs may be fed to the algorithms during training phase (Stankovic et al 2012;Ganguly 2008;Bishop 2006). The performance of a classifier depends on its ability to classify the unseen data based on the learned model and is more generally known as its generalization ability (Nguyen et al 2008;Yeung et al 2007;Steinwart and Christmann 2008;Tax and Duin 2004). Depending on the type of model learned during the training phase, these techniques can be divided into two types:…”
Section: Classification Based Outlier and Event Detection For Wsns Dementioning
confidence: 99%
“…Reduction Sampling [9][10][11][12][13][14][15][16][17][18] Kernel Matrix [24][25][26][27][28][29] Optimization [5][6][7][8] Fast Inference [2,[30][31][32][33] post-processing pre-processing efficient if the number of support vectors is low. Also note that under mild assumptions, SVDD is equivalent to ν-SVM [20].…”
Section: Fast Trainingmentioning
confidence: 99%
“…This is the category of methods mentioned in our introduction [9][10][11][12][13][14][15][16][17][18]. A second type reduces the size of the Kernel matrix, e.g., by approximation [24][25][26][27]. Examples are the Nystrm-method [28] and choosing random Fourier features [29].…”
Section: A Categorizationmentioning
confidence: 99%
“…In such cases, they can log only aggregated data or an approximation of aggregated data and still get a good estimate of the required statistics. Approximation provides statistically sound estimates of metrics that are useful to machine-learning analyses such as PCA (principal component analysis) and SVM (support vector machine 14 ). These techniques are critical in networked or large-scale distributed systems, where collecting even a single number from each component carries a heavy performance cost.…”
mentioning
confidence: 99%