Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP 2020
DOI: 10.18653/v1/2020.blackboxnlp-1.20
|View full text |Cite
|
Sign up to set email alerts
|

Exploring Neural Entity Representations for Semantic Information

Abstract: Neural methods for embedding entities are typically extrinsically evaluated on downstream tasks and, more recently, intrinsically using probing tasks. Downstream task-based comparisons are often difficult to interpret due to differences in task structure, while probing task evaluations often look at only a few attributes and models. We address both of these issues by evaluating a diverse set of eight neural entity embedding methods on a set of simple probing tasks, demonstrating which methods are able to remem… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
5
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 37 publications
0
5
0
Order By: Relevance
“…This progress has been accompanied by the creation of entity-driven datasets for tasks such as language modeling [1,37,59], question answering [32,34,42,71,87], fact checking [4,55,73] and information extraction [85,89], to name a few. Yet, recent findings [18,24,41,64,70,76] suggest that entity representation and identification (i.e., identifying the correct entity that match a given text) are among the main challenges that should be solved to further increase performance on such datasets. We believe that TempEL can contribute to addressing these challenges by: (i) encouraging research on devising more robust methods to creating entity representations that are invariant to temporal changes; and (ii) improving entity identification for non-trivial scenarios involving ambiguous and uncommon mentions (e.g., linked to overshadowed entities as defined above).…”
Section: Entity-driven Datasetsmentioning
confidence: 99%
“…This progress has been accompanied by the creation of entity-driven datasets for tasks such as language modeling [1,37,59], question answering [32,34,42,71,87], fact checking [4,55,73] and information extraction [85,89], to name a few. Yet, recent findings [18,24,41,64,70,76] suggest that entity representation and identification (i.e., identifying the correct entity that match a given text) are among the main challenges that should be solved to further increase performance on such datasets. We believe that TempEL can contribute to addressing these challenges by: (i) encouraging research on devising more robust methods to creating entity representations that are invariant to temporal changes; and (ii) improving entity identification for non-trivial scenarios involving ambiguous and uncommon mentions (e.g., linked to overshadowed entities as defined above).…”
Section: Entity-driven Datasetsmentioning
confidence: 99%
“…This progress has been accompanied by the creation of entity-driven datasets for tasks such as language modeling [238][239][240], question answering [241][242][243][244][245], fact checking [16,17,246] and information extraction [4,48], to name a few. Yet, recent findings [21,[247][248][249][250][251] identifying the correct entity that match a given text) are among the main challenges that should be solved to further increase performance on such datasets. We believe that TempEL can contribute to addressing these challenges by: (i) encouraging research on devising more robust methods to creating entity representations that are invariant to temporal changes; and (ii) improving entity identification for non-trivial scenarios involving ambiguous and uncommon mentions (e.g., linked to overshadowed entities as defined above).…”
Section: Entity-driven Datasetsmentioning
confidence: 99%
“…A variety of embedding algorithms have been developed for learning representations of domain concepts and real-world entities from text, including weakly-supervised methods requiring only a terminology (Newman-Griffis et al, 2018); methods using pre-trained NER models for noisy annotation (De Vine et al, 2014;Chen et al, 2020); and methods leveraging explicit annotations of concept mentions (as in Wikipedia) (Yamada et al, 2020). 1 These algorithms capture valuable patterns about concept types and relationships that can inform corpus analysis (Runge and Hovy, 2020).…”
Section: From Words To Domain Conceptsmentioning
confidence: 99%