2019
DOI: 10.48550/arxiv.1911.02855
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Dice Loss for Data-imbalanced NLP Tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
46
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 39 publications
(46 citation statements)
references
References 33 publications
0
46
0
Order By: Relevance
“…Most of the samples do not belong to any specific class (non-concept or non-named-entity). As showed in [15], this imbalance can rise to a ratio of 168:1. One option to handle this imbalance is to over or under-sample the data [33], or selecting features based on their importance for the minority class [31].…”
Section: Related Workmentioning
confidence: 89%
See 1 more Smart Citation
“…Most of the samples do not belong to any specific class (non-concept or non-named-entity). As showed in [15], this imbalance can rise to a ratio of 168:1. One option to handle this imbalance is to over or under-sample the data [33], or selecting features based on their importance for the minority class [31].…”
Section: Related Workmentioning
confidence: 89%
“…One option to handle this imbalance is to over or under-sample the data [33], or selecting features based on their importance for the minority class [31]. Other approaches to overcoming the negative consequences of these extreme imbalanced classes are to use different loss functions [15] or to adopt different weights for the distinct classes [22].…”
Section: Related Workmentioning
confidence: 99%
“…For example, binary cross-entropy (BCE) loss, which is also referred to as log loss, often serves as the default loss metric for binary classification tasks [134][135][136]. However, it often performs poorly when data are imbalanced [134,137,138]. As a result, it may be more appropriate to apply class weighting, or weighted binary cross-entropy (WBCE), when classes are imbalanced [134,135].…”
Section: Note On Loss Metricsmentioning
confidence: 99%
“…As a result, it may be more appropriate to apply class weighting, or weighted binary cross-entropy (WBCE), when classes are imbalanced [134,135]. Alternatively, 1-F1 Score or 1-Dice, generally referred to as Dice loss, is more robust to data imbalance than BCE [137][138][139][140]. It is even possible to combine multiple loss functions, such as BCE and Dice loss with equal or different weighting applied to each loss component [134].…”
Section: Note On Loss Metricsmentioning
confidence: 99%
“…We introduce two main modifications to the original PUnet model. First, we replace the reconstruction loss by a Dice loss (Li et al, 2019), more suited to our unbalanced segmentation problem (see Section 3). We weight the Dice loss so that a wrong prediction on a blend as more importance than a wrong prediction on the background, which dominates the image.…”
Section: Modelmentioning
confidence: 99%