Proceedings of the 55th Annual Meeting of the Association For Computational Linguistics (Volume 2: Short Papers) 2017
DOI: 10.18653/v1/p17-2003
|View full text |Cite
|
Sign up to set email alerts
|

Lexical Features in Coreference Resolution: To be Used With Caution

Abstract: Lexical features are a major source of information in state-of-the-art coreference resolvers.Lexical features implicitly model some of the linguistic phenomena at a fine granularity level. They are especially useful for representing the context of mentions. In this paper we investigate a drawback of using many lexical features in state-of-the-art coreference resolvers. We show that if coreference resolvers mainly rely on lexical features, they can hardly generalize to unseen domains. Furthermore, we show that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
12
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
3
2

Relationship

1
9

Authors

Journals

citations
Cited by 24 publications
(12 citation statements)
references
References 11 publications
0
12
0
Order By: Relevance
“…Generally, models may learn to memorize artifact and biases rather than truly learning (Gururangan et al, 2018;Moosavi and Strube, 2017;Agrawal et al, 2016), e.g., from political individuals often leaning towards one side of the truth spectrum. Additionally, language models have been shown to implicitly store world knowledge (Roberts et al, 2020), which in principle could enhance the aforementioned biases.…”
Section: Related Workmentioning
confidence: 99%
“…Generally, models may learn to memorize artifact and biases rather than truly learning (Gururangan et al, 2018;Moosavi and Strube, 2017;Agrawal et al, 2016), e.g., from political individuals often leaning towards one side of the truth spectrum. Additionally, language models have been shown to implicitly store world knowledge (Roberts et al, 2020), which in principle could enhance the aforementioned biases.…”
Section: Related Workmentioning
confidence: 99%
“…They model these properties as continuous scores associated to each mention and bucketized for evaluation. Lexical overlap has also been mentioned in Coreference Resolution (Moosavi and Strube, 2017) where coreferent mentions tend to co-occur in the test and train sets. In this line of works, the impact of lexical overlap is measured either by separating performance depending on the property of mentions (seen or unseen) or with outof-domain evaluation with a test set from a different dataset with lower lexical overlap with the train set.…”
Section: Related Workmentioning
confidence: 99%
“…In general, coreference resolution methods are divided into rulebased methods, machine learning-based (statistical), and deep learning-based groups. In rule-based methods [21][22][23][24][25][26][27][28], a collection of rules are handwritten by experts. These rules are implemented in an orderly manner to specify co-referents in the text.…”
Section: Related Workmentioning
confidence: 99%