Medical Imaging 2021: Imaging Informatics for Healthcare, Research, and Applications 2021
DOI: 10.1117/12.2582226
|View full text |Cite
|
Sign up to set email alerts
|

Annotation quality vs. quantity for deep-learned medical image segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 0 publications
0
7
0
Order By: Relevance
“… 37 Contrary to the previous studies, Wesemeyer et al. 38 demonstrated a trade-off between quality and quantity for deep learning segmentation performance. As we illustrate in our study, expert consistency may be better than nonexpert consistency for some radiotherapy-related ROIs, particularly for H&N imaging.…”
Section: Discussionmentioning
confidence: 85%
See 1 more Smart Citation
“… 37 Contrary to the previous studies, Wesemeyer et al. 38 demonstrated a trade-off between quality and quantity for deep learning segmentation performance. As we illustrate in our study, expert consistency may be better than nonexpert consistency for some radiotherapy-related ROIs, particularly for H&N imaging.…”
Section: Discussionmentioning
confidence: 85%
“…36 Additionally, the authors show no statistical difference between the variability of experts vs. novices in several of the annotations of interest, results which are echoed in a recent study investigating crowdsourcing for liver tumor segmentation where the quality of annotations was not statistically significantly different in four distinct groups. 37 Contrary to the previous studies, Wesemeyer et al 38 demonstrated a trade-off between quality and quantity for deep learning segmentation performance. As we illustrate in our study, expert consistency may be better than nonexpert consistency for some radiotherapy-related ROIs, particularly for H&N imaging.…”
Section: Nonexpert Staple Bootstrap Experimentsmentioning
confidence: 85%
“…Additionally, the authors show average DSC for expert-expert and novice-expert segmentations show that no statistical difference is found between the variability of annotators in several of the annotations of interest, results which are echoed in a recent study investigating crowdsourcing for liver tumor segmentation where the quality of annotations was not statistically significantly different four distinct groups [35]. Contrary to the previous studies, Wesemeyer et al demonstrate a tradeoff between quality and quantity for deep learning segmentation performance [36]. As we demonstrate in our study, expert consistency may be better than non-expert consistency for some radiotherapy-related ROIs, particularly for H&N imaging.…”
Section: Discussionmentioning
confidence: 98%
“…However, for auto-segmentation algorithms to be clinically useful, their input data (training data) should reflect high-quality "gold-standard" annotations. While research has been performed on the impact of interobserver variability and segmentation quality for auto-segmentation training [8][9][10][11] , it remains unclear how "gold-standard" segmentations should be defined and generated. One common approach, consensus segmentation generation, seeks to crowdsource multiple segmentations from different annotators to generate a high-quality ground-truth segmentation.…”
Section: Background and Summarymentioning
confidence: 99%