2016
DOI: 10.1007/978-3-319-46454-1_24
|View full text |Cite
|
Sign up to set email alerts
|

SPICE: Semantic Propositional Image Caption Evaluation

Abstract: There is considerable interest in the task of automatically generating image captions. However, evaluation is challenging. Existing automatic evaluation metrics are primarily sensitive to n-gram overlap, which is neither necessary nor sufficient for the task of simulating human judgment. We hypothesize that semantic propositional content is an important component of human caption evaluation, and propose a new automated caption evaluation metric defined over scene graphs coined SPICE. Extensive evaluations acro… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

6
1,195
3

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 1,369 publications
(1,204 citation statements)
references
References 38 publications
(81 reference statements)
6
1,195
3
Order By: Relevance
“…BLEU-1, BLEU-2, BLEU-3 and BLEU-4) based on the n-gram method of determining string/sentence similarity. It is also equipped with other evaluation metrics such as, METEOR [31], ROUGE-L [32], and SPICE [33]. The detailed results using all of the above metrics for the evaluation of the IAPR TC-12 dataset are shown in Table 3.…”
Section: Discussionmentioning
confidence: 99%
“…BLEU-1, BLEU-2, BLEU-3 and BLEU-4) based on the n-gram method of determining string/sentence similarity. It is also equipped with other evaluation metrics such as, METEOR [31], ROUGE-L [32], and SPICE [33]. The detailed results using all of the above metrics for the evaluation of the IAPR TC-12 dataset are shown in Table 3.…”
Section: Discussionmentioning
confidence: 99%
“…Following standard practice (Anderson et al, 2016;Elliott and Keller, 2014), we compared with Text in red is extra information, while text in green is missing information. sim(x, y) is the average similarity between machine-identified and true error vectors over image regions.…”
Section: Methodsmentioning
confidence: 99%
“…19 The automatic evaluation measures include BLEU-1,-2,-3,-4 (Papineni et al 2002), METEOR (Denkowski and Lavie 2014), ROUGE-L (Lin 2004), and CIDEr . We also use the recently proposed evaluation measure SPICE (Anderson et al 2016), which aims to compare the semantic content of two descriptions, by matching the information contained in dependency parse trees for both descriptions. While we report all measures for the final evaluation in the LSMDC (Sect.…”
Section: Automatic Metricsmentioning
confidence: 99%