2017
DOI: 10.1016/j.cviu.2017.05.017
|View full text |Cite
|
Sign up to set email alerts
|

Vision-language integration using constrained local semantic features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2017
2017
2020
2020

Publication Types

Select...
2
1
1

Relationship

2
2

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…Three kinds of categories are added: specific [3], [4], [37], [59] (e.g. rottweiler), generic [28], [31], [49] (e.g. dog) and noisy [25], [51].…”
Section: Learning One Network On a Modified Spmentioning
confidence: 99%
See 3 more Smart Citations
“…Three kinds of categories are added: specific [3], [4], [37], [59] (e.g. rottweiler), generic [28], [31], [49] (e.g. dog) and noisy [25], [51].…”
Section: Learning One Network On a Modified Spmentioning
confidence: 99%
“…We propose a new method that takes advantage of the principle of the re-training of neural networks on the same problems, and thus does not need more data [3], [31], [49], nor increasing the network capacity [1], [44], [47], [54]. Our approach relates to the work of [58] who proposes an extensive study of the effect of different self-training methods (i.e., re-training a neural network on the same problem it was trained originally).…”
Section: Focused Self Fine-tunningmentioning
confidence: 99%
See 2 more Smart Citations
“…Multimodal corpora. Many corpora provide images with associated textual content, in particular for the tasks of automatic image annotation (Young et al, 2014;Ginsca et al, 2015), cross-media retrieval (Karpathy and Fei-Fei, 2015;Tran et al, 2016a), image-sentence matching (Hodosh et al, 2013;Ordonez et al, 2011), text illustration (Feng and Lapata, 2010;Chami et al, 2017) and cross-media classification (Tran et al, 2016b;Tamaazousti et al, 2017). Most corpora used in this context consist in images with captions from Flickr (Ordonez et al, 2011;Hodosh et al, 2013;Young et al, 2014) or using Amazon's Mechanical Turk (Rashtchian et al, 2010;Lin et al, 2014).…”
Section: Related Workmentioning
confidence: 99%