Proceedings of the 29th ACM International Conference on Multimedia 2021
DOI: 10.1145/3474085.3476968
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal Relation Extraction with Efficient Graph Alignment

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
23
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 29 publications
(24 citation statements)
references
References 30 publications
1
23
0
Order By: Relevance
“…Detailed statistics are shown in Table 3. For multimodal RE, we evaluate on MNRE [59], a manually-labeled dataset for multimodal neural relation extraction, where the texts and image posts are crawled from Twitter. For multimodal NER, we conduct experiments on public Twitter dataset Twitter-2017 [25], which mainly include multimodal user posts published on Twitter during 2016-2017.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Detailed statistics are shown in Table 3. For multimodal RE, we evaluate on MNRE [59], a manually-labeled dataset for multimodal neural relation extraction, where the texts and image posts are crawled from Twitter. For multimodal NER, we conduct experiments on public Twitter dataset Twitter-2017 [25], which mainly include multimodal user posts published on Twitter during 2016-2017.…”
Section: Methodsmentioning
confidence: 99%
“…Besides, Sergieh et al [33] and Wang et al [47] jointly encode and fuse the visual and structural knowledge for multimodal link prediction through simple concatenation and autoencoder, respectively. On the other hand, Zheng et al [59] present an efficient modality alignment strategy based on scene graph for the MRE task. Zhang et al [55] fuse regional image features and textual features with extra co-attention layers for the MNER task.…”
Section: Superman Returnsmentioning
confidence: 99%
See 3 more Smart Citations