Advances in Soft Computing
DOI: 10.1007/3-540-31662-0_33
|View full text |Cite
|
Sign up to set email alerts
|

Taming the Curse of Dimensionality in Kernels and Novelty Detection

Abstract: Abstract. The curse of dimensionality is a well known but not entirely well-understood phenomena. Too much data, in terms of the number of input variables, is not always a good thing. This is especially true when the problem involves unsupervised learning or supervised learning with unbalanced data (many negative observations but minimal positive observations). This paper addresses two issues involving high dimensional data: The first issue explores the behavior of kernels in high dimensional data. It is shown… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
36
0

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 42 publications
(39 citation statements)
references
References 20 publications
(9 reference statements)
2
36
0
Order By: Relevance
“…The deterioration of the performance of Kernel_MEPSO is in accordance with (Evangelista et al, 2006), which indicates that higher dimensionality confounds the process of kernel based learning especially in presence of unbalanced classes. The curse of dimensionality may be overcome in Kernel based clustering by utilizing the subspace models.…”
Section: Scalability Of the Kernel_mepso Algorithmmentioning
confidence: 52%
“…The deterioration of the performance of Kernel_MEPSO is in accordance with (Evangelista et al, 2006), which indicates that higher dimensionality confounds the process of kernel based learning especially in presence of unbalanced classes. The curse of dimensionality may be overcome in Kernel based clustering by utilizing the subspace models.…”
Section: Scalability Of the Kernel_mepso Algorithmmentioning
confidence: 52%
“…In complex problems, the number of parameters that need to be set within the cognitive model is large and cannot be readily identified even with large training data sets. Because of the range of possible actions, these problems suffer from the "curse of dimensionality" where even sophisticated machine learning approaches and parameter tuning alone will not suffice (Evangelista et al, 2006;Bengio et al, 2005). In these problems, the novel hybrid approach we present-using machine learning Combining Psychological Models with Machine Learning 17 algorithms as the base of the solution, but adding features from the cognitive models -creates significant improvements over both base approaches, often by large amounts.…”
Section: Discussionmentioning
confidence: 99%
“…These domains suffer from a phenomenon often known as the "curse of dimensionality" (Evangelista, Embrechts, & Szymanski, 2006;Bengio, Delalleau, & Roux, 2005). The curse of dimensionality is, unfortunately, particularly evident in many real-world situations, as people can potentially act within a very large set of actions (a high dimension of possible actions), and they often do not consistently choose the same actions.…”
Section: Introductionmentioning
confidence: 99%
“…In addition, we use random forest (RF) both for feature selection and to benchmark the performance of variants of kNN. RF is a commonly used robust classifier which can handle large amount of features without being affected by the curse of dimensionality due to inherent internal randomised distribution of data among decision trees [86].…”
Section: B the Learnersmentioning
confidence: 99%
“…The nearest neighbour learners are particularly sensitive to the curse of dimensionality [39] especially with small k values. High dimensionality is not a problem with RF due to distribution of randomised subsets of features and samples among decision trees [86]. Therefore, it is reasonable to ask that if we do feature selection before classification, is it likely that we will get better results?…”
Section: Tile-rfmentioning
confidence: 99%