2019
DOI: 10.48550/arxiv.1906.07413
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss

Abstract: Deep learning algorithms can fare poorly when the training dataset suffers from heavy class-imbalance but the testing criterion requires good generalization on less frequent classes. We design two novel methods to improve performance in such scenarios. First, we propose a theoretically-principled label-distribution-aware margin (LDAM) loss motivated by minimizing a margin-based generalization bound. This loss replaces the standard cross-entropy objective during training and can be applied with prior strategies… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
120
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
1
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 90 publications
(120 citation statements)
references
References 27 publications
0
120
0
Order By: Relevance
“…FashionMNIST [], CIFAR-10 and CIFAR-100 [] with controllable degrees of data imbalance. Follow the setting in [3], the magnitude of the long-tailed imbalance declines exponentially with class size. The imbalance ratio ρ represents the ratio between the samples of the most and least frequent classes.…”
Section: Experiments Setupsmentioning
confidence: 99%
See 4 more Smart Citations
“…FashionMNIST [], CIFAR-10 and CIFAR-100 [] with controllable degrees of data imbalance. Follow the setting in [3], the magnitude of the long-tailed imbalance declines exponentially with class size. The imbalance ratio ρ represents the ratio between the samples of the most and least frequent classes.…”
Section: Experiments Setupsmentioning
confidence: 99%
“…[9,1,2]. It is usually the case that real-world data exhibit a imbalanced distribution, and highly skewed data can adversely affect the effectiveness of machine learning [10,3]. Re-sampling [11] and re-weighting [12,10,13] are traditional methods towards addressing imbalanced data.…”
Section: Learning From Imbalanced Datamentioning
confidence: 99%
See 3 more Smart Citations