2016
DOI: 10.1007/978-3-319-46487-9_19
|View full text |Cite
|
Sign up to set email alerts
|

The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition

Abstract: Current approaches for fine-grained recognition do the following: First, recruit experts to annotate a dataset of images, optionally also collecting more structured data in the form of part annotations and bounding boxes. Second, train a model utilizing this data. Toward the goal of solving fine-grained recognition, we introduce an alternative approach, leveraging free, noisy data from the web and simple, generic methods of recognition. This approach has benefits in both performance and scalability. We demonst… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

5
256
0
1

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 282 publications
(271 citation statements)
references
References 59 publications
5
256
0
1
Order By: Relevance
“…Text-mining module: When this was not used, the overall accuracy dropped as the irrelevant training labels brought noises. However, the performance did not degrade substantially, showing that our model is able to tolerate noisy labels to a certain degree [18]. We also found training with the relevant + uncertain labels was better than using relevant labels only, which is because most uncertain labels are radiologists' inferences that are very likely to be true, especially if we only consider the lesion's appearance.…”
Section: Methodsmentioning
confidence: 78%
See 1 more Smart Citation
“…Text-mining module: When this was not used, the overall accuracy dropped as the irrelevant training labels brought noises. However, the performance did not degrade substantially, showing that our model is able to tolerate noisy labels to a certain degree [18]. We also found training with the relevant + uncertain labels was better than using relevant labels only, which is because most uncertain labels are radiologists' inferences that are very likely to be true, especially if we only consider the lesion's appearance.…”
Section: Methodsmentioning
confidence: 78%
“…Noisy and incomplete training labels often exist in datasets mined from the web [7], which is similar to our labels mined from reports. Strategies to handle them include data filtering [18], noise-robust losses [31], noise modeling [25], finding reliable negative samples [20], and so on. We use a text-mining module to filter noisy positive labels and leverage label relation to find reliable negative labels.…”
Section: Related Workmentioning
confidence: 99%
“…Besides visual information, researchers have been using additional information such as pose [5], attributes [25] and text description [17]. Data augmentation and transfer learning have also been studied [14,7].…”
Section: Related Workmentioning
confidence: 99%
“…Countering the view that noise needs to be carefully modelled, [17] shows that a large quantity of noisy data can be effectively used to increase performance on a fine-grained classification task. From a similar perspective, [18] attempts to find the limits of weakly supervised data by pre-training with 3.5 billion images for classification and object detection tasks.…”
Section: Related Workmentioning
confidence: 99%