Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.519
|View full text |Cite
|
Sign up to set email alerts
|

Scalable Zero-shot Entity Linking with Dense Entity Retrieval

Abstract: This paper introduces a conceptually simple, scalable, and highly effective BERT-based entity linking model, along with an extensive evaluation of its accuracy-speed trade-off. We present a two-stage zero-shot linking algorithm, where each entity is defined only by a short textual description. The first stage does retrieval in a dense space defined by a bi-encoder that independently embeds the mention context and the entity descriptions. Each candidate is then re-ranked with a crossencoder, that concatenates t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
371
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 244 publications
(434 citation statements)
references
References 17 publications
2
371
0
Order By: Relevance
“…Entities in each claim are identified with BLINK (Wu et al, 2019), a model trained on Wikipedia data that links each entity to its nearest Wikipedia page. BLINK combines a bi-encoder (Urbanek et al, 2019;) that identifies candidates with a cross-encoder that models the interaction between mention context and entity descriptions.…”
Section: Entity Briefsmentioning
confidence: 99%
“…Entities in each claim are identified with BLINK (Wu et al, 2019), a model trained on Wikipedia data that links each entity to its nearest Wikipedia page. BLINK combines a bi-encoder (Urbanek et al, 2019;) that identifies candidates with a cross-encoder that models the interaction between mention context and entity descriptions.…”
Section: Entity Briefsmentioning
confidence: 99%
“…We observed improvements on most frequency buckets compared to DE R@1, which suggests that the model's few-shot capability can be improved by cross-lingual reading-comprehension. This also offers an initial multilingual validation of a similar two-step BERT-based approach recently introduced in a monolingual setting by (Wu et al, 2019), and provides a strong baseline for future work.…”
Section: Outcomementioning
confidence: 92%
“…Basing entity representations on features of their Wikipedia pages has been a common approach in EL (e.g. Sil and Florian, 2016;Francis-Landau et al, 2016;Gillick et al, 2019;Wu et al, 2019), but we will need to generalize this to include multiple Wikipedia pages with possibly redundant features in many languages.…”
Section: Mel With Wikidata and Wikipediamentioning
confidence: 99%
See 1 more Smart Citation
“…By using active sampling, we minimize labeling efforts. TrainX uses transfer learning by leveraging Bi-Encoders (Gillick et al, 2019;Wu et al, 2019;Logeswaran et al, 2019;Humeau et al, 2020) for disambiguation and a kNN-index to retrieve candidate entities within milliseconds. We mitigate issues caused by sparse training data by using zero-shot optimized techniques that can generalize beyond the labels seen in training.…”
Section: Contributionmentioning
confidence: 99%