Findings of the Association for Computational Linguistics: EMNLP 2020 2020
DOI: 10.18653/v1/2020.findings-emnlp.228
|View full text |Cite
|
Sign up to set email alerts
|

Zero-shot Entity Linking with Efficient Long Range Sequence Modeling

Abstract: This paper considers the problem of zero-shot entity linking, in which a link in the test time may not present in training. Following the prevailing BERT-based research efforts, we find a simple yet effective way is to expand the long-range sequence modeling. Unlike many previous methods, our method does not require expensive pre-training of BERT with long position embeddings. Instead, we propose an efficient position embeddings initialization method called Embedding-repeat, which initializes larger position e… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(27 citation statements)
references
References 15 publications
0
10
0
Order By: Relevance
“…BLINK (Wu et al, 2020) proposes a bi-encoder to encode the descriptions and enhance the bi-encoder by distilling the knowledge from the cross-encoder. Yao et al (2020) repeats the position embedding to solve the long-range modeling problem in entity descriptions. Zhang and Stratos (2021) demonstrates that hard negatives can enhance the contrast when training an EL model.…”
Section: Related Workmentioning
confidence: 99%
“…BLINK (Wu et al, 2020) proposes a bi-encoder to encode the descriptions and enhance the bi-encoder by distilling the knowledge from the cross-encoder. Yao et al (2020) repeats the position embedding to solve the long-range modeling problem in entity descriptions. Zhang and Stratos (2021) demonstrates that hard negatives can enhance the contrast when training an EL model.…”
Section: Related Workmentioning
confidence: 99%
“…As an encoder BERT model is selected. While state-of-the-art models in Zero-shot EL (Logeswaran et al, 2019;Yao et al, 2020) focus on the CR phase, Wu et al (2020) are the only to propose a different to traditional IR approach for CG. Our focus is to further push the boundaries of the CG phase and set a higher performance threshold to CR and EL overall.…”
Section: State-of-the-art Cg Modelsmentioning
confidence: 99%
“…Most EL systems consist of two subsystems: Candidate Generation (CG), where for each entity mention the system detects entities related to the mention and document, and Candidate Ranking (CR) where the system chooses the most probable entity link among the found candidates. Most state-of-the-art models (Logeswaran et al, 2019;Li et al, 2020;Yao et al, 2020) rely on traditional frequency-based CG and focus on building robust candidate rankers using cross-encoders to jointly encode mention and entity candidate descriptions. However, the memory-intensive CR phase depends on the set of candidates provided by CG.…”
Section: Introductionmentioning
confidence: 99%
“…BLINK (Wu et al, 2020) proposes a bi-encoder to encode the descriptions and enhance the bi-encoder by distilling the knowledge from the cross-encoder. Yao et al (2020) repeats the position embedding to solve the long-range modeling problem in entity descriptions. Zhang and Stratos (2021) demonstrates that hard negatives can enhance the contrast when training an EL model.…”
Section: Related Workmentioning
confidence: 99%