2012
DOI: 10.1007/978-3-642-27355-1_5
|View full text |Cite
|
Sign up to set email alerts
|

Combining Image-Level and Segment-Level Models for Automatic Annotation

Abstract: Abstract. For the task of assigning labels to an image to summarize its contents, many early attempts use segment-level information and try to determine which parts of the images correspond to which labels. Best performing methods use global image similarity and nearest neighbor techniques to transfer labels from training images to test images. However, global methods cannot localize the labels in the images, unlike segment-level methods. Also, they cannot take advantage of training images that are only locall… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2012
2012
2015
2015

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 22 publications
0
2
0
Order By: Relevance
“…In addition, one image is represented by many visual feature vectors, resulting in high computational cost (see table 1). [45] More characteristic layout [53] Do not capture spatial information [54] Sensitive to intensity variations and distortion [37] Fail to narrow down the semantic gap due to their limited descriptive power based on objects [37] Do not have good performance [54] Weak in characterizing the internal content of image [49] Not recommended for multiple complex object images. Limited interpretability [53] …”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, one image is represented by many visual feature vectors, resulting in high computational cost (see table 1). [45] More characteristic layout [53] Do not capture spatial information [54] Sensitive to intensity variations and distortion [37] Fail to narrow down the semantic gap due to their limited descriptive power based on objects [37] Do not have good performance [54] Weak in characterizing the internal content of image [49] Not recommended for multiple complex object images. Limited interpretability [53] …”
Section: Discussionmentioning
confidence: 99%
“…More precise, discriminating and explicit [53] [52] Spatial information [1] Good for search of specific object [52] [45] Improve classification [48] More flexible and Compositional character [53] Good generalization potential [53] High computational cost [49] High number of matches for a simple query [52] Need additional processing (e.g. segmentation) Not recommended when searching complex information [52] Produce unsatisfactory accuracy [49] …”
Section: Localmentioning
confidence: 99%