Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL) 2019
DOI: 10.18653/v1/k19-1063
|View full text |Cite
|
Sign up to set email alerts
|

Investigating Entity Knowledge in BERT with Simple Neural End-To-End Entity Linking

Abstract: arXiv:2003.05473v1 [cs.CL]

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
109
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 87 publications
(112 citation statements)
references
References 20 publications
3
109
0
Order By: Relevance
“…Entity linking (EL) is the task of detecting entity spans in a text and linking them to the underlying entity ID. While there are recent advances in fully end-to-end EL (Broscheit, 2019), the task is typically broken down into three steps: (1) detecting spans that are potential entity spans, (2) generating sets of candidate entities for these spans, (3) selecting the correct candidate for each span.…”
Section: Entity Linkingmentioning
confidence: 99%
See 1 more Smart Citation
“…Entity linking (EL) is the task of detecting entity spans in a text and linking them to the underlying entity ID. While there are recent advances in fully end-to-end EL (Broscheit, 2019), the task is typically broken down into three steps: (1) detecting spans that are potential entity spans, (2) generating sets of candidate entities for these spans, (3) selecting the correct candidate for each span.…”
Section: Entity Linkingmentioning
confidence: 99%
“…To ensure coverage of the necessary entities, we include all gold entities and all generator candidates in the entity vocabulary L Ent , even if they fall under the Wikipedia2Vec link threshold (see Section 3.3). While this is based on the unrealistic assumption that we know the contents of the test set in advance, it is necessary for comparability with Peters et al ( 2019), Kolitsas et al (2018) and Broscheit (2019), who also design their entity vocabulary around the data. See Appendix for more details on data and preprocessing.…”
Section: Finetuning We Finetune E-bert-mlm On the Training Set To Minimizementioning
confidence: 99%
“…Several works (Radford et al, 2019b;Keskar et al, 2019) show the remarkable fluency and gram-matical correctness of text decoded from modern LMs. Additionally, recent works (Petroni et al, 2019;Logan et al, 2019;Broscheit, 2019;Roberts et al, 2020) demonstrate that beyond general linguistic capabilities, language models can also pick up factual knowledge present in the training data. However, it is unclear if LMs are able to convey such knowledge at decoding time when producing long sequences-do they generate fluent, grammatical but "babbler-level" text or can they produce utterances that reflect factual world knowledge?…”
Section: Introductionmentioning
confidence: 99%
“…For short text, the features provided by the context are limited, and the features cannot be well represented, which leads to some models not fully learning the features of the text. Therefore, some scholars proposed pre-training a model to address this problem [19], connecting the text and description as the input of BERT, and then classifying the vector output by [CLS] position together with the start position and end position vectors of the candidate entities. However, because the features extracted by the BERT model are relatively broad and noisy, we believe performance on this task can be further improved.…”
Section: Related Workmentioning
confidence: 99%
“…Finally, the fused semantic features are input into the fully connected layer with a sigmoid activation function for classification, as shown in formulas (18) and (19).…”
Section: Fusion Layermentioning
confidence: 99%