2022
DOI: 10.1609/aaai.v36i3.20258
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Logit Adjustment Loss for Long-Tailed Visual Recognition

Abstract: Data in the real world tends to exhibit a long-tailed label distribution, which poses great challenges for the training of neural networks in visual recognition. Existing methods tackle this problem mainly from the perspective of data quantity, i.e., the number of samples in each class. To be specific, they pay more attention to tail classes, like applying larger adjustments to the logit. However, in the training process, the quantity and difficulty of data are two intertwined and equally crucial problems. For… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 23 publications
(17 citation statements)
references
References 34 publications
(67 reference statements)
0
17
0
Order By: Relevance
“…We randomly select ten images per class from training data to construct metadata. Several classical and advanced robust loss functions and data augmentation approaches that are mainly designed for long-tailed classifications are compared, including Classbalanced CE loss [31], Class-balanced Focal loss, LDAM [23], LDAM-DRW [23], ISDA [7], LA [24], ALA [33], LPL [32], MixUp [35], and RISDA [9]. Besides, De-confound-TDE [34], which uses causal intervention in training and counterfactual reasoning in inference, is also involved in our comparison.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…We randomly select ten images per class from training data to construct metadata. Several classical and advanced robust loss functions and data augmentation approaches that are mainly designed for long-tailed classifications are compared, including Classbalanced CE loss [31], Class-balanced Focal loss, LDAM [23], LDAM-DRW [23], ISDA [7], LA [24], ALA [33], LPL [32], MixUp [35], and RISDA [9]. Besides, De-confound-TDE [34], which uses causal intervention in training and counterfactual reasoning in inference, is also involved in our comparison.…”
Section: Methodsmentioning
confidence: 99%
“…Methods designed for long-tailed classification including Class-balanced CE loss [31], OLTR [40], LDAM [23], LDAM-DRW [23], LA [24], ALA [33], RISDA [9], Meta-classweight [36], and MetaSAug [8] are compared. Results.…”
Section: Methodsmentioning
confidence: 99%
“…To cope with this problem, GCL [14] introduces Gaussian clouds into the logit adjustment process, and adaptively sets the cloud size according to the sample size of each class. Recently, Zhao et al [46] reveal that previous logit adjustment techniques primarily focus on the sample quantity of each class, while ignoring the difficulty of samples. Thus, Zhao et al [46] propose to prevent the over-optimization on easy samples of tail classes, while highlighting the training on difficult samples of head classes.…”
Section: Balanced Lossmentioning
confidence: 99%
“…Recently, Zhao et al [46] reveal that previous logit adjustment techniques primarily focus on the sample quantity of each class, while ignoring the difficulty of samples. Thus, Zhao et al [46] propose to prevent the over-optimization on easy samples of tail classes, while highlighting the training on difficult samples of head classes.…”
Section: Balanced Lossmentioning
confidence: 99%
“…Also, it is feasible for various places to look similar, particularly within similar environments, an error called perceptual aliasing [4]. Primarily applied for computer vision (CV) tasks, Convolution Neural Network (CNN) based model has been pioneered in the VPR fields during the past few years, obtaining remarkable achievements on different data sets [5].…”
Section: Introductionmentioning
confidence: 99%