2021
DOI: 10.1007/978-3-030-73197-7_35
|View full text |Cite
|
Sign up to set email alerts
|

Attention-Based Multimodal Entity Linking with High-Quality Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
27
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 19 publications
(27 citation statements)
references
References 23 publications
0
27
0
Order By: Relevance
“…To alleviate the long-term dependency problem and extract worthy information, some DL based EL systems leverage BERT [83] to encode the entity description in order to learn the entity embedding. Specifically, several EL works [67], [68], [69], [77] took the entity description as a sequence of words as input to BERT. Most of these works [68], [69], [77] inserted a special start token [CLS] at the beginning of the input sequence and the output of the last layer at this start token produced by the Transformer encoder is regarded as the vector representation of the input sequence.…”
Section: Entity Descriptionmentioning
confidence: 99%
See 4 more Smart Citations
“…To alleviate the long-term dependency problem and extract worthy information, some DL based EL systems leverage BERT [83] to encode the entity description in order to learn the entity embedding. Specifically, several EL works [67], [68], [69], [77] took the entity description as a sequence of words as input to BERT. Most of these works [68], [69], [77] inserted a special start token [CLS] at the beginning of the input sequence and the output of the last layer at this start token produced by the Transformer encoder is regarded as the vector representation of the input sequence.…”
Section: Entity Descriptionmentioning
confidence: 99%
“…Specifically, several EL works [67], [68], [69], [77] took the entity description as a sequence of words as input to BERT. Most of these works [68], [69], [77] inserted a special start token [CLS] at the beginning of the input sequence and the output of the last layer at this start token produced by the Transformer encoder is regarded as the vector representation of the input sequence. Fang et al [67] obtained the entity embedding via average-pooling over the hidden states of all description tokens in the last BERT layer.…”
Section: Entity Descriptionmentioning
confidence: 99%
See 3 more Smart Citations