Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.680
|View full text |Cite
|
Sign up to set email alerts
|

Temporally-Informed Analysis of Named Entity Recognition

Abstract: Natural language processing models often have to make predictions on text data that evolves over time as a result of changes in language use or the information described in the text. However, evaluation results on existing data sets are seldom reported by taking the timestamp of the document into account. We analyze and propose methods that make better use of temporally-diverse training data, with a focus on the task of named entity recognition. To support these experiments, we introduce a novel data set of En… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
33
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1
1

Relationship

2
8

Authors

Journals

citations
Cited by 38 publications
(40 citation statements)
references
References 38 publications
2
33
0
Order By: Relevance
“…Similar to scenario 1, the F1 scores of the models trained on instances selected based on their trend scores are always higher than random sampling F1 scores. In addition, scenario 2, on average, works better than scenario 1, which is consistent with Rijhwani and Preotiuc-Pietro (2020). However, this setting requires the data available from all years from the very beginning.…”
Section: Resultssupporting
confidence: 72%
“…Similar to scenario 1, the F1 scores of the models trained on instances selected based on their trend scores are always higher than random sampling F1 scores. In addition, scenario 2, on average, works better than scenario 1, which is consistent with Rijhwani and Preotiuc-Pietro (2020). However, this setting requires the data available from all years from the very beginning.…”
Section: Resultssupporting
confidence: 72%
“…Since we have time stamps for two of our datasets we study these in greater detail. For similar studies of temporal drift, see Lukes and Søgaard (2018); Rijhwani and Preotiuc-Pietro (2020). The training data sizes are comparable (1.63-1.76M), the publisher distributions (AFP, APW, CNA, NYT or XIN) are also similar.…”
Section: A4 Computing Adversarial Splitsmentioning
confidence: 87%
“…In addition to research on improving the performance of the NER model, other experimental setups have been proposed for this task. These include domain adaptation, where a model trained on data from a source domain is used to tag data from a different target domain (Guo et al, 2009;Greenberg et al, 2018;Wang et al, 2020), temporal drift, where a model is tested on data from future time intervals (Derczynski et al, 2016;Rijhwani and Preotiuc-Pietro, 2020), cross-lingual modelling where models trained in one language are adapted to other languages (Tsai et al, 2016;Ni et al, 2017;Xie et al, 2018), identifying nested entities Lu and Roth, 2015) or high-precision NER models (Arora et al, 2019).…”
Section: Related Workmentioning
confidence: 99%