2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2016
DOI: 10.1109/cvpr.2016.29
|View full text |Cite
|
Sign up to set email alerts
|

What Value Do Explicit High Level Concepts Have in Vision to Language Problems?

Abstract: Much recent progress in Vision-to-Language (V2L) problems has been achieved through a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). This approach does not explicitly represent high-level semantic concepts, but rather seeks to progress directly from image features to text. In this paper we investigate whether this direct approach succeeds due to, or despite, the fact that it avoids the explicit representation of high-level information. We propose a method of incorpora… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
279
2
2

Year Published

2017
2017
2019
2019

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 395 publications
(285 citation statements)
references
References 38 publications
(86 reference statements)
0
279
2
2
Order By: Relevance
“…In [10,42], the authors introduced a framework which utilizes a pre-trained CNN as an encoder to extract image features, followed by an RNN as a decoder to generate image descriptions. This model was further improved by incorporating high-level semantic attribute information [44,49] or regularizing the RNN decoder [6]. To distill the salient objects or important regions from an image, different kinds of attention mechanisms were integrated into the captioning framework to exam the relevant image regions when generating sentences [2,26,45,47,50].…”
Section: Related Workmentioning
confidence: 99%
“…In [10,42], the authors introduced a framework which utilizes a pre-trained CNN as an encoder to extract image features, followed by an RNN as a decoder to generate image descriptions. This model was further improved by incorporating high-level semantic attribute information [44,49] or regularizing the RNN decoder [6]. To distill the salient objects or important regions from an image, different kinds of attention mechanisms were integrated into the captioning framework to exam the relevant image regions when generating sentences [2,26,45,47,50].…”
Section: Related Workmentioning
confidence: 99%
“…Automatic image captioning has drawn great attention in recent years [8][9][10][11][12][13][14][15][16][17][18][19][20][21]. Karpathy and Li [8] proposed a system to provide natural language descriptions for image regions.…”
Section: Image Description Generationmentioning
confidence: 99%
“…Recent work seems to suggest that, in the endto-end learning framework, using posterior distributions over a refined set of object classes (relevant to captions) performs better than using lower level dense image representations (Wu et al, 2016;You et al, 2016). Vinyals et al (2016) note that using a better image network (a network that performs better on the image classification task) results in improvements in the generated captions.…”
Section: Studying Visual Representationsmentioning
confidence: 99%