2015 IEEE International Conference on Computer Vision (ICCV) 2015
DOI: 10.1109/iccv.2015.525
|View full text |Cite
|
Sign up to set email alerts
|

Love Thy Neighbors: Image Annotation by Exploiting Image Metadata

Abstract: Some images that are difficult to recognize on their own may become more clear in the context of a neighborhood of related images with similar social-network metadata. We build on this intuition to improve multilabel image annotation. Our model uses image metadata nonparametrically to generate neighborhoods of related images using Jaccard similarities, then uses a deep neural network to blend visual information from the image and its neighbors. Prior work typically models image metadata parametrically; in cont… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
117
0
1

Year Published

2016
2016
2021
2021

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 102 publications
(118 citation statements)
references
References 41 publications
(62 reference statements)
0
117
0
1
Order By: Relevance
“…Such networks learn very promising features for many tasks, such as image classification [15] and object detection [23]. Global CNN-based image features have been used for image annotation too [13]; however, some recent work [10] [27] learns local features for detected bounding boxes, so as to extract more discriminative object-centric features rather than from background. The second direction focuses on exploring and exploiting tag correlations.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Such networks learn very promising features for many tasks, such as image classification [15] and object detection [23]. Global CNN-based image features have been used for image annotation too [13]; however, some recent work [10] [27] learns local features for detected bounding boxes, so as to extract more discriminative object-centric features rather than from background. The second direction focuses on exploring and exploiting tag correlations.…”
Section: Related Workmentioning
confidence: 99%
“…As such, image annotation is treated as a multi-label learning problem, where tag correlations play a key role. Most common tag correlations involve tag-level smoothness [30,32] (i.e., the prediction scores of two semantically similar tags should be similar in the same image), image-level smoothness [13,30,32,20] (i.e., visually similar images have similar tags), low rank assumption [2] (i.e., the whole tag space is spanned by a lower-dimensional space), and semantic hierarchy [30,25] (i.e. parent tags in a hierarchy are as probable as their children).…”
Section: Related Workmentioning
confidence: 99%
“…In addition to tags and images, this group of works exploits user information, motivated from varied perspectives. User information ranges from the simplest user identities [Li et al 2009b], to tagging preferences [Sawant et al 2010], to user reliability [Ginsca et al 2014], to image group memberships [Johnson et al 2015]. With the hypothesis that a specific tag chosen by many users to label visually similar images is more likely to be relevant with respect to the visual content, Li et al [2009b] utilize user identity to ensure that learning examples come from distinct users.…”
Section: Media For Tag Relevancementioning
confidence: 99%
“…In McParlane et al [2013b], time-constrained tag co-occurrence statistics are considered to refine the output of visual classifiers for tag assignment. In their follow-up work [McParlane et al 2013a], location-constrained tag co-occurrence computed from images Johnson et al [2015], social network metadata such as image groups membership or contacts of users is employed to resolve ambiguity in visual appearance. Comparing the three groups, tag + image appears to be the mainstream, as evidenced by the imbalanced distribution in Table I.…”
Section: Media For Tag Relevancementioning
confidence: 99%
“…In the field of image processing and pattern recognition, the research effort has transformed from image recognition to image annotation step by step [1]. According to machine learning and pattern recognition algorithm, the traditional image recognition can finish the work of feature extraction and classification.…”
Section: Introductionmentioning
confidence: 99%