2018
DOI: 10.1007/978-3-030-05587-5_23
|View full text |Cite
|
Sign up to set email alerts
|

Improving SNR and Reducing Training Time of Classifiers in Large Datasets via Kernel Averaging

Abstract: Kernel methods are of growing importance in neuroscience research. As an elegant extension of linear methods, they are able to model complex non-linear relationships. However, since the kernel matrix grows with data size, the training of classifiers is computationally demanding in large datasets. Here, a technique developed for linear classifiers is extended to kernel methods: In linearly separable data, replacing sets of instances by their averages improves signal-to-noise ratio (SNR) and reduces data size. I… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 20 publications
0
3
0
Order By: Relevance
“…For linear classifiers, this approach has been explored by (Cichy et al, 2015;Cichy and Pantazis, 2017). Recently, it has been generalized to non-linear kernel methods (Treder, 2018). Either approach can be used in the toolbox by adding the operation average_samples or average_kernel to the preprocessing pipeline.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…For linear classifiers, this approach has been explored by (Cichy et al, 2015;Cichy and Pantazis, 2017). Recently, it has been generalized to non-linear kernel methods (Treder, 2018). Either approach can be used in the toolbox by adding the operation average_samples or average_kernel to the preprocessing pipeline.…”
Section: Discussionmentioning
confidence: 99%
“…That is, they are performed on the training data first and subsequently applied to the test data using parameters estimated from the training data (Lemm et al, 2011;Varoquaux et al, 2017). Currently implemented functions include PCA, sample averaging (Cichy and Pantazis, 2017), kernel averaging (Treder, 2018), and under-/oversampling for unbalanced data. Preprocessing pipelines are defined by adding the cfg.preprocess parameter.…”
Section: Preprocessingmentioning
confidence: 99%
“…The most often used classifier, random forest (RF), builds an ensemble classifier by combining many CART trees [33]. RF builds several decision trees using a random selection of training datasets and parameters.…”
Section: Classification Algorithmsmentioning
confidence: 99%