2020
DOI: 10.1007/978-3-030-58568-6_34
|View full text |Cite
|
Sign up to set email alerts
|

Finding It at Another Side: A Viewpoint-Adapted Matching Encoder for Change Captioning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
26
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 35 publications
(29 citation statements)
references
References 22 publications
0
26
0
Order By: Relevance
“…To validate the generalization ability of the proposed method, we conduct the experiments on a recent published Spot-the-Diff dataset, where the image pairs are mostly well aligned and their is no viewpoint change. We compare with eight SOTA methods and most of them cannot consider handling viewpoint changes: DDLA (Jhamtani and Berg-Kirkpatrick, 2018), DDUA (Park et al, 2019), SDCM (Oluwasanmi et al, 2019a), FCC (Oluwasanmi et al, 2019b), static rel-att / dyanmic rel-att (Tan et al, 2019), and M-VAM / M-VAM+RAF (Shi et al, 2020).…”
Section: Results On Spot-the-diffmentioning
confidence: 99%
See 3 more Smart Citations
“…To validate the generalization ability of the proposed method, we conduct the experiments on a recent published Spot-the-Diff dataset, where the image pairs are mostly well aligned and their is no viewpoint change. We compare with eight SOTA methods and most of them cannot consider handling viewpoint changes: DDLA (Jhamtani and Berg-Kirkpatrick, 2018), DDUA (Park et al, 2019), SDCM (Oluwasanmi et al, 2019a), FCC (Oluwasanmi et al, 2019b), static rel-att / dyanmic rel-att (Tan et al, 2019), and M-VAM / M-VAM+RAF (Shi et al, 2020).…”
Section: Results On Spot-the-diffmentioning
confidence: 99%
“…However, it is built upon an ideal situation by assuming there are no distractors (illumination/viewpoint change) between a pair of images. To make this task more close to our dynamic world, Park et al and Shi et al (Park et al, 2019;Shi et al, 2020) both aimed to address change captioning in the existence of distractors. On one hand, Park et al directly concatenated the coarse feature difference with the image pair to operate spatial attention to localize the change.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations
“…Hence, feature shift between two unaligned images will adversely affect the learning of difference representation. To make this task more practical, recent works (Park et al, 2019;Shi et al, 2020) proposed to address change captioning in the presence of viewpoint changes. Despite the progress, there are some limitations for the above state-of-the-art methods when modeling the difference representation.…”
Section: Introductionmentioning
confidence: 99%