2021
DOI: 10.1080/02664763.2021.1911965
|View full text |Cite
|
Sign up to set email alerts
|

Detection of outliers in high-dimensional data using nu-support vector regression

Abstract: Support Vector Regression (SVR) is gaining in popularity in the detection of outliers and classification problems in high-dimensional data (HDD) as this technique does not require the data to be of full rank. In real application, most of the data are of high dimensional. Classification of high-dimensional data is needed in applied sciences, in particular, as it is important to discriminate cancerous cells from non-cancerous cells. It is also imperative that outliers are identified before constructing a model o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 29 publications
0
2
0
Order By: Relevance
“…These data points are embedded vectors of climate events. Since there are 1–10% anomalies in a dataset 14 , we conduct the experiments to validate the performance of the proposed model by selecting 1–10% anomalies. For example, we assume that 10% of data points are selected as outliers in each dataset.…”
Section: Resultsmentioning
confidence: 99%
“…These data points are embedded vectors of climate events. Since there are 1–10% anomalies in a dataset 14 , we conduct the experiments to validate the performance of the proposed model by selecting 1–10% anomalies. For example, we assume that 10% of data points are selected as outliers in each dataset.…”
Section: Resultsmentioning
confidence: 99%
“…A dataset may contain multiple outliers, posing challenges in detecting and addressing the masking and swamping effects [7]. Various methods have been employed for multiple outlier detection, including the fully efficient one-step procedure (GY) proposed by Gervini and Yohai (2002) [8], the least trimmed squares (LTS) [9], and the MM-estimators [10].…”
Section: Introductionmentioning
confidence: 99%
“…Support Vector Regression (SVR), effective in high-dimensional spaces and robust to outliers, demands careful selection of kernel and parameters due to its computational intensity [ 25 ]. Decision Tree Regression, with its capability to handle non-linearity and interactions, is visually interpretable but prone to overfitting and sensitive to small variations in data.…”
Section: Introductionmentioning
confidence: 99%