2018
DOI: 10.1109/tnnls.2017.2755595
|View full text |Cite
|
Sign up to set email alerts
|

A Distance-Based Weighted Undersampling Scheme for Support Vector Machines and its Application to Imbalanced Classification

Abstract: A support vector machine (SVM) plays a prominent role in classic machine learning, especially classification and regression. Through its structural risk minimization, it has enjoyed a good reputation in effectively reducing overfitting, avoiding dimensional disaster, and not falling into local minima. Nevertheless, existing SVMs do not perform well when facing class imbalance and large-scale samples. Undersampling is a plausible alternative to solve imbalanced problems in some way, but suffers from soaring com… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 135 publications
(37 citation statements)
references
References 32 publications
0
35
0
Order By: Relevance
“…The G-mean (denoted as GM) evaluates class-wise sensitivity and indicates the balanced classification performances on the majority and minority classes. The micro average scheme M AUC [15] is defined as the area under the curve metric. As for the task of object detection, we utilize the Average Precision (AP) (IoU=[.50:.05:.95]), AP .50 (IoU=.50), and AP .75 (IoU=.75) as performance evaluation metrics.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…The G-mean (denoted as GM) evaluates class-wise sensitivity and indicates the balanced classification performances on the majority and minority classes. The micro average scheme M AUC [15] is defined as the area under the curve metric. As for the task of object detection, we utilize the Average Precision (AP) (IoU=[.50:.05:.95]), AP .50 (IoU=.50), and AP .75 (IoU=.75) as performance evaluation metrics.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…1) Sampling-Based Methods: Sampling-based methods attempt to handle the class imbalance problem at the data level, i.e., improving the data preprocessing technique. Specifically, these methods aim to balance the distribution of the original training set by over-sampling the minority classes [40]- [43], under-sampling the majority classes [7], [44], [45], or both.…”
Section: A Class Imbalancementioning
confidence: 99%
“…As for the area under the curve (AUC) metric in the classification problem, we follow the micro average scheme M AUC of the definition as in [7]. Similar to the form of F -measure and G-mean, it integrates the weighted average of all labels:…”
Section: B Experimental Settings 1) Training/testing Set Partitionmentioning
confidence: 99%
See 1 more Smart Citation
“…The algorithms based on frequency can be realized at a lower computation cost by reducing the dimension of the frequency vectors [56]. SVM is sensitive to noises and outliers [57][58][59][60]. FSVM is based on fuzzy theory to reduce the influence of noises or outliers on the classification hyperplane [61][62][63][64][65][66].…”
Section: Previous Workmentioning
confidence: 99%