2021
DOI: 10.1016/j.imavis.2021.104316
|View full text |Cite
|
Sign up to set email alerts
|

Aligning vision-language for graph inference in visual dialog

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…However, ML techniques without dimensionality reduction perform better with low-dimensional datasets Pei et al 31 2022 This work proposes scene graph semantic inference for cross-modal image and text matching. Using visual and textual scene graphs and graph convolutional networks, the technique analyses the local semantic correlations of inter-modal object relationships Jiang et al 32 2021 This paper proposes visual dialogue for establishing semantic relationships between visual and written content. Aligning visual and textual knowledge reduce the gap between modalities.…”
Section: Introductionmentioning
confidence: 99%
“…However, ML techniques without dimensionality reduction perform better with low-dimensional datasets Pei et al 31 2022 This work proposes scene graph semantic inference for cross-modal image and text matching. Using visual and textual scene graphs and graph convolutional networks, the technique analyses the local semantic correlations of inter-modal object relationships Jiang et al 32 2021 This paper proposes visual dialogue for establishing semantic relationships between visual and written content. Aligning visual and textual knowledge reduce the gap between modalities.…”
Section: Introductionmentioning
confidence: 99%