Procedings of the British Machine Vision Conference 2011 2011
DOI: 10.5244/c.25.29
|View full text |Cite
|
Sign up to set email alerts
|

A Large-Scale Database of Images and Captions for Automatic Face Naming

Abstract: We present a large scale database of images and captions, designed for supporting research on how to use captioned images from the Web for training visual classifiers. It consists of more than 125,000 images of celebrities from different fields downloaded from the Web. Each image is associated to its original text caption, extracted from the html page the image comes from. We coin it FAN-Large, for Face And Names Large scale database. Its size and deliberate high level of noise makes it to our knowledge the la… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
21
0

Year Published

2012
2012
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(21 citation statements)
references
References 16 publications
(34 reference statements)
0
21
0
Order By: Relevance
“…2) Results on the real-world datasets : For performance evaluation, we follow [37] to take the accuracy and precision as two criteria. The accuracy is the percentage of correctly annotated faces (also including the correctly annotated faces whose groundtruth name is the "null" name) over all faces, while the precision is the percentage of correctly annotated faces over the faces which are annotated as real names (i.e., we do not consider the faces annotated as the "null" class by a face naming method).…”
Section: ) Results On the Synthetic Datasetmentioning
confidence: 99%
“…2) Results on the real-world datasets : For performance evaluation, we follow [37] to take the accuracy and precision as two criteria. The accuracy is the percentage of correctly annotated faces (also including the correctly annotated faces whose groundtruth name is the "null" name) over all faces, while the precision is the percentage of correctly annotated faces over the faces which are annotated as real names (i.e., we do not consider the faces annotated as the "null" class by a face naming method).…”
Section: ) Results On the Synthetic Datasetmentioning
confidence: 99%
“…Our models are trained over web images queried from Bing Image search engine for the same names. All the data preprocessing and the feature extraction flow follow the same line of [41], that is owned from [12]. However, [41] trains the models and evaluates the results at the same collection.…”
Section: Learning Facesmentioning
confidence: 99%
“…We use FAN-large [41] face data-set for testing our method in face recognition problem. We use Easy and Hard subsets with the names accommodating more than 100 images (to have fair testing results).…”
Section: Learning Facesmentioning
confidence: 99%
“…news documents (6.9M words). The attachment probabilities (see (24)) were estimated from the same corpus. We tuned the caption length parameter on the development set using a range of ½5; 14 tokens for the word-based model and ½2; 5 phrases for the phrase-based model.…”
Section: Parameter Tuningmentioning
confidence: 99%
“…Examples include associating names mentioned in the captions to faces depicted in news images (e.g., [23], [24]), verbs to body poses [25], and learning models for recognizing objects [26] and their relative importance [27].…”
Section: Related Workmentioning
confidence: 99%