2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.656
|View full text |Cite
|
Sign up to set email alerts
|

Diverse Image Annotation

Abstract: In this work, we study a new image annotation task called diverse image annotation (DIA)

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
1

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
4

Relationship

2
7

Authors

Journals

citations
Cited by 29 publications
(18 citation statements)
references
References 30 publications
0
17
1
Order By: Relevance
“…The reason is that D 2 IA-GAN may include more irrelevant tags, as the random noise combined with the image feature not only brings in diversity, but also uncertainty. Note that due to the randomness of sampling, the results of single subset by DIA presented here are slightly different with those reported in [23].…”
Section: Quantitative Resultscontrasting
confidence: 86%
See 1 more Smart Citation
“…The reason is that D 2 IA-GAN may include more irrelevant tags, as the random noise combined with the image feature not only brings in diversity, but also uncertainty. Note that due to the randomness of sampling, the results of single subset by DIA presented here are slightly different with those reported in [23].…”
Section: Quantitative Resultscontrasting
confidence: 86%
“…They are also different in the training process, which will be reviewed in the Section 4. Besides, in DIA [23], 'diverse/diversity' refers to the semantic difference between tags in the same tag subset, to which we use the word 'distinct/distinctiveness' for the same meaning in this work. We use 'diverse/diversity' to indicate the semantic difference between multiple tag subsets for the same image.…”
Section: Related Workmentioning
confidence: 99%
“…Dense captioning [9] aims to identify all the salient regions in an image and describe each with a caption. Diverse image annotation [44] focuses on describing as much of the image as possible with a limited number of tags. Entity-aware captioning [25] employs hashtags as additional input.…”
Section: Descriptive Captioningmentioning
confidence: 99%
“…We evaluate accuracy at k = 1 and k = 10, which measure how often the first ranked hashtag is in the groundtruth and how often at least one of the 10 highest ranked hashtags is in the groundtruth respectively. A desired feature of a tagging system is the ability to infer diverse and distinct tags [42,43]. In order to measure the variety of tags predicted by the models, we measure the percentage of all the test tags predicted at least once in the whole test set (%pred) and the percentage of all the test tags correctly predicted at least once (%cpred), considering the top 10 tags predicted for each image.…”
Section: Image Taggingmentioning
confidence: 99%