2021
DOI: 10.48550/arxiv.2112.05958
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

You Only Need End-to-End Training for Long-Tailed Recognition

Abstract: The generalization gap on the long-tailed data sets is largely owing to most categories only occupying a few training samples. Decoupled training achieves better performance by training backbone and classifier separately. What causes the poorer performance of end-to-end model training (e.g., logits margin-based methods)? In this work, we identify a key factor that affects the learning of the classifier: the channel-correlated features with low entropy before inputting into the classifier. From the perspective … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 27 publications
0
2
0
Order By: Relevance
“…We conducted an ablative study to analyze which of the commonly used loss function CE and L f was more appropriate for stage one; • We compared our full pipeline with common methods in the literature proposed for handling class imbalance, namely cost-sensitive loss (CS) [38], class-balanced loss by effective number of classes (CB) [23], focal loss (FL) [22], label-distribution-aware margin loss (LDAM) [41], influence-balanced Loss (IB) [60], bag of tricks (BAGs) [50] and decoupled training [21]; • We compared our approach with prior works developing CAD systems for SL classification; • We analyzed the best performance achieved with our pipelines.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…We conducted an ablative study to analyze which of the commonly used loss function CE and L f was more appropriate for stage one; • We compared our full pipeline with common methods in the literature proposed for handling class imbalance, namely cost-sensitive loss (CS) [38], class-balanced loss by effective number of classes (CB) [23], focal loss (FL) [22], label-distribution-aware margin loss (LDAM) [41], influence-balanced Loss (IB) [60], bag of tricks (BAGs) [50] and decoupled training [21]; • We compared our approach with prior works developing CAD systems for SL classification; • We analyzed the best performance achieved with our pipelines.…”
Section: Resultsmentioning
confidence: 99%
“…Current existing approaches dealing with class imbalance can be subdivided into three approaches [17]: data processing, cost-sensitive weighting, and decoupling methods. The decoupled training seems to achieve better performance than the reweighting methods [21]. In general, a decoupled training involves a two-stage pipeline that learns representations under the imbalance dataset at the first stage, then rebalances the classifier with a frozen representation at the second stage.…”
Section: Introductionmentioning
confidence: 99%