Proceedings of the 2011 SIAM International Conference on Data Mining 2011
DOI: 10.1137/1.9781611972818.17
|View full text |Cite
|
Sign up to set email alerts
|

A Quadratic Mean based Supervised Learning Model for Managing Data Skewness

Abstract: In this paper, we study the problem of data skewness. A data set is skewed/imbalanced if its dependent variable is asymmetrically distributed. Dealing with skewed data sets has been identified as one of the ten most challenging problems in data mining research. We address the problem of class skewness for supervised learning models which are based on optimizing a regularized empirical risk function. These include both classification and regression models for discrete and continuous dependent variables. Classic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Year Published

2013
2013
2023
2023

Publication Types

Select...
4
4
2

Relationship

4
6

Authors

Journals

citations
Cited by 21 publications
(7 citation statements)
references
References 26 publications
0
7
0
Order By: Relevance
“…The function (4) is the risk function derived from the maximum likelihood estimation of θ. The arithmetic (5) and quadratic (6) risk functions are intended to perform better under class imbalance conditions by computing the per-class risk [14]. The arithmetic risk takes the sum of the per-class values, whereas the quadratic risk uses the root of the sum of the squared values.…”
Section: Multinomial Logistic Regressionmentioning
confidence: 99%
“…The function (4) is the risk function derived from the maximum likelihood estimation of θ. The arithmetic (5) and quadratic (6) risk functions are intended to perform better under class imbalance conditions by computing the per-class risk [14]. The arithmetic risk takes the sum of the per-class values, whereas the quadratic risk uses the root of the sum of the squared values.…”
Section: Multinomial Logistic Regressionmentioning
confidence: 99%
“…[8], and are likely to enable the future self-driving cars [9]. However, recent studies have shown that DNNs tend to deliver a wrong prediction with high confidence if the input data are slightly perturbed [10][11] [12][13] [14]. These perturbations are often imperceptible to human vision system, but their effects on DNNs are catastrophic [15].…”
Section: Targeted Attention Attack On Deep Learningmentioning
confidence: 99%
“…If the input data points were not connected in the form of a graph, the problem could have been formulated as either semi-supervised outlier detection [12] or supervised classification from imbalanced data [6,16,5,19,15,11] depending on whether degree of supervision is low or high. The existing literature, however, appears to be quite slim when it comes to address the problem of semi-supervised graph label propagation for rare labels.…”
Section: Graph Label Propagation -Prior Artmentioning
confidence: 99%